Sunday, March 12

  • Registration is open from 8:00 AM - 4:00 PM each day.

  • Get hands-on training in the areas of (listed below) by attending a Short Course.

    Learn more >

    • Introduction to Deep Learning with PyTorch (9:00 AM)

    • Bayesian Methods in A/B Testing (9:00 AM)

    • Adapting Pretrained Models for Document Classification (12:30 PM)

    • Introduction to the Graph Neural Networks and Their Generalizations (12:30 PM)

  • Learn more >

    The DSCO Datathon provides a platform for high school and undergraduate data science enthusiasts to learn, apply, and hone their data science skills through a particular problem statement and data set that will be presented by Amazon Web Services at the start of the competition.

  • Henry Humadi and Shan Wang, Conference Co-Chairs

    Jeff Hamrick, Senior Director of the Data Institute

  • AI Accelerating Science: Neural Operators for Learning on Function Spaces

    In this talk, Anima will present exciting developments in the use of AI for scientific simulations. This includes diverse domains such as weather and climate modeling, carbon capture and storage, fluid dynamics etc. We have developed principled neural operator framework that enables zero-shot generalization beyond the training discretization or resolution. They yield 4-5 orders of magnitude speedups over current numerical methods. They learn mappings between function spaces that makes them ideal for capturing multi-scale processes.

  • Join us for a reception where you can catch up with industry colleagues, make new friends, and create valuable connections. Music, refreshments, and hors d'oeuvres will be provided.

    Make sure to add this valuable networking experience to your conference schedule!

Monday, March 13

  • Registration is open 8:00 AM - 4:00 PM each day.

  • Online controlled experiments (i.e., A/B tests) have seen a meteoric rise in prominence in the last decade. Generally speaking, such experiments seek to test and improve internet-based products and services using user-generated data to determine what works and what doesn’t. A/B tests have become an indispensable tool for major tech companies when it comes to maximizing revenue and optimizing the user experience, with some companies running hundreds of experiments engaging millions of users each day.

    In this online setting, and with a culture of testing as many ideas as possible, as quickly as possible, novel practical issues and modern challenges abound.

    This session brings together researchers and practitioners from both industry and academia to discuss such modern challenges and their work on them.

    Session Chair: Nathaniel Stevens

    Session Sponsor: University of San Francisco Master of Science in Applied Economics

    Speakers & Abstracts:
    Lo-Hua Yuan, Airbnb: How Airbnb Learns from Product Releases When We Can't A/B Test

    Airbnb runs thousands of A/B tests (i.e., randomized controlled experiments) each year, but one notable exception is its biannual major product release. In these major releases, many new features are bundled together and launched globally at the same time for publicity purposes, making standard A/B testing infeasible. Yet, we still need to understand the impacts and causal dynamics behind these global launches, so that we can decide how to prioritize feature iterations post-launch. In this talk, I discuss how we built a measurement framework for Airbnb’s global product releases by casting the problem as an interrupted time series (ITS) design, where we have pre- and post-intervention time series data for only a single treated unit and no comparable controls. The cornerstone of the ITS design is counterfactual time series forecasting: i) Build a model using pre-intervention time series to forecast counterfactual post-intervention trajectories of what we would expect to see in the absence of the intervention; ii) Compare these estimated counterfactual trends to the trends we actually observe post-intervention to estimate time-varying impacts and their uncertainty. I describe several Bayesian regression ITS techniques that we implemented, the tooling we built to scale accessibility and reusability of the ITS workflow, and practical learnings from challenges that we tackled.

    Cindy Zhang, Pinterest: Long-term Holdouts as Ground Truth Measurements
    At Pinterest we run thousands of online experiments to make decisions on features. Since 2022, we have been running team-level holdout experiments, where we hold back all new features developed by a team in a quarter from certain users. With individual experiments for each of the features and holdout experiments side by side, we observed numerous examples of an effect in the opposite direction to that expected from the winner's curse. The aggregated result from individual experiments showed an underestimate rather than an overestimate as the winner’s curse states, which has led to some changes in how we think about setting goals as a team and demonstrated the importance of holdouts. In this talk, we’ll discuss the potential causes of this phenomenon, the best practices for setting up team-level holdout experiments and our other learnings from the process.

    Jiannan Lu, formerly Microsoft ExP: Towards Trustworthy, Efficient and Private A/B tests: Learnings and Insights
    Modern A/B testing on-device poses various statistical challenges, from treatment deployment, to data collection, processing and analytics. In this talk, we share our learnings in addressing these challenges, in particular on data-privacy friendly aggregations, and de-biasing methods for treatment non-compliance. Our goal is to not only discuss exploratory solutions, but more importantly, highlight open questions that might be of interest to both researchers and practitioners.

    Nicholas Larsen, NCSU: HODOR: A Two-Stage Hold-Out Design for Online Controlled Experiments on Networks
    A/B tests are standard tools for estimating the average treatment effect in online controlled experiments (OCEs). The majority of OCE theory relies on the Stable Unit Treatment Value Assumption, which presumes the response of individual users depends only on the assigned treatment, not the treatments of others. Violations of this assumption occur when users are subjected to network interference, a common phenomenon in social media platforms. Standard methods for estimating the average treatment effect typically ignore network effects and produce heavily biased results. Additionally, unobserved user covariates, such as offline information or variables hidden due to privacy restrictions, that influence user response and network structure also bias current estimators of the average treatment effect. In this paper, we demonstrate that network-influential lurking variables can heavily bias popular network clustering-based methods, thereby making them unreliable. To address this problem, we propose a two-stage design and estimation technique called HODOR (Hold-Out Design for Online Randomized experiments). We show that HODOR is unbiased for the average treatment effect, has minimizable variance, and provides reliable estimation even when the underlying network is partially-unknown or uncertain.

  • Session Chair: Steve Devlin, University of San Francisco

    Speakers:

    Justin Jacobs, Squared2020, NBA: A Brief Tour into Computer Vision in Sports

    With the advances in deep learning and computer processing, difficult tasks such as classifying objects in video have become much more achievable for the sports analyst. In this talk, we will take a look at ways to perform object detection in video and discuss the pros and cons of various types of methods such as optical flow, Yolo, and part affinity fields. Specifically, we will break down some known packages available to the public and identify how to extract data to begin asking more challenging questions.

    Jake Toffler, New York Mets: Model Benchmarking in a Competitive Industry Abstract

    How do you know how your models compare to the competition when everyone wants to keep their edge a secret?

    Thomas Treloar, Hillsdale College: Global and Iterative Ranking Models Based on Network Diffusion

    In this talk we discuss families of ranking models based on network diffusion processes. The framework of a diffusion process unites these various methods and provides a strong intuition into how and why they differ. Included in this framework are global methods (i.e. Markov, Colley, Massey, Bradley-Terry rankings) and iterative methods (i.e. Elo and Local Markov rankings).

    David Uminsky, University of Chicago: NBA Lineup Attribution: A Fourier Transform Approach

    Basketball analytics has seen real growth in methodology and practice in line with the explosive increase in the use of data in professional sports over the past decade. Nonetheless, basketball data has unique challenges and hard questions that remain open for rigorous inquiry. These include: Do winning teams need a "big 3"?Am I a good player, or am I only a good player because I'm on the court with Lebron James? Generally, understanding a player’s contribution to winning basketball individually as well as a member of subgroups on the floor is difficult. In this talk we will present a new, Fourier transform based mathematical tool that was designed to orthogonally break down the player attribution problem. We will demonstrate the tool and findings using lineup level play-by-play data from the 2015-2016 NBA season.

  • Session Chair: Jeff Hamrick

    Speakers:

    OLIVER ZEIGERMANN, Open Knowledge: Resilient Machine Learning

    Resilience is known as the ability to adapt to difficult or unexpected situations. Such a phenomenon also exists in the field of machine learning, where we have to deal with adversarial attacks, out-of distribution robustness, drift, and unwanted bias. Additionally, model stability must be ensured during retraining and post-training. Measures include monitoring machine learning models, detecting outliers, and running with fallbacks and/or multiple models as an ensemble. Clever model selection for production can also go a long way. In this talk, I explain the phenomena mentioned and address the appropriate measures based on an example in the field of computer vision, which poses additional challenges.

    PHIL MUI, Salesforce: An Embedding Analysis on the Impact of Diversity Awareness on Diversity of Ideas

    Diversity in its various forms is often touted as vital for progress and social justice. In particular, many colleges are expected to foster "diversity" in admission, student body, and faculty hiring. etc. Has the recent emphasis on diversity across colleges led to a corresponding increase in the diversity of ideas being discussed?

    We analyze student opinion writings in college newspapers from the past decade, using word vectors to measure the range of ideas discussed. We then compare this data with the trend in the number of times diversity-related words are mentioned in the same newspapers.

    Our main finding is that explicit mentions of diversity have definitely increased across colleges in the past decade. However, the diversity of ideas has often decreased in most colleges during this time. This seems to indicate that while colleges are encouraging more discourse about diversity, the actual space of ideas being discussed in school newspaper opinion sections has unfortunately been shrinking.

    While colleges proclaim their desire to promote diversity on their campuses, our research suggests that this desire may not always be reflected in the diversity of ideas present in college newspapers. To ensure that diverse voices are not drowned out, it is important that we continue to examine and address this issue.

    KICHO YU, University of Southern California: Flexible and Robust Real-Time Intrusion Detection Systems to Network Dynamics

    Deep learning-based intrusion detection systems have advanced due to their technological innovations such as high accuracy, automation, and scalability to develop an effective network intrusion detection system (NIDS). However, most of the previous research has focused on model generation through intensive analysis of feature engineering instead of considering real environments. They have limitations to applying the previous methods for a real network environment to detect real-time network attacks. In this paper, we propose a new flexible and robust NIDS based on Recurrent Neural Network (RNN) with a multi-classifier to generate a detection model in real-time. The proposed system adaptively and intelligently adjusts the generated model with given system parameters that can be used as security parameters to defend against the attacker’s obfuscation techniques in real-time. In the experimental results, the proposed system detects network attacks with a high accuracy and high-speed model upgrade in real-time while showing robustness under an attack.

    CONNOR GIBBS, Colorado State University: ECoHeN: A hypothesis testing framework to extract communities from heterogeneous networks

    Community discovery is a process of identifying assortative communities in a network: collections of nodes which are densely connected within but sparsely connected to the rest of the network. While community discovery has been extensively studied, there are few techniques available for heterogeneous networks that contain different types of nodes and possibly different connectivity patterns between the node types. In this talk, we introduce a framework called ECoHeN to extract communities from a heterogeneous network in a statistically meaningful way. ECoHeN uses a heterogeneous configuration model as a reference distribution to identify communities that are significantly more densely connected than expected given the node types and connectivity of its membership. The ECoHeN algorithm extracts communities one at a time using a dynamic set of iterative updating rules, is guaranteed to converge, and imposes no constraints on the type composition of extracted communities. To our knowledge, ECoHeN is the first method that can distinguish and identify both homogeneous and heterogeneous, possibly overlapping, community structure in a network. We demonstrate the utility of ECoHeN in the context of a popular political blogs network to identify collections of blogs that reference one another more than expected considering the ideology of its members.

  • Join mentors and mentees for a sponsored lunch where you can enjoy good food and meaningful conversation.

  • DJ Patil will sit down with Jeff Hamrick (Data Institute) to talk about his time as our country’s first Chief Data Scientist and much more.

  • The proliferation of AI has exposed a new area of risk to data privacy and security. Although data may be secured when at rest, ML models may divulge private traces of the data when included in a training set. Data in transit may be secured cryptographically, yet may unintentionally convey sensitive information about individuals to unintended parties through dissemination by the recipient.

    Zero trust for hosted ML services presents a challenge of sharing data that may be insecure “at inference”. Additionally, efforts to improve model privacy may be at odds with efforts to track model provenance of large language models that might be used in disinformation campaigns.

    In this session, invited speakers confront challenges to keep data secure and private when used to train ML models.

    Session Chair:

    Hyrum Anderson

    Confirmed Speakers:

    DANIEL ZIELASKI, Salesforce: The Importance of AI to Business, Marketing, and Digital Products
    Artificial intelligence (AI) is one of the most rapidly advancing technologies of our time, with the potential to revolutionize the way businesses operate. As a subset of data science, AI can provide companies with valuable insights into customer behavior, preferences, and trends, enabling them to make informed decisions that drive growth and success. In this presentation, we will explore the importance of AI in marketing, business, and digital product use cases. In addition, we will examine how AI can be used in the development of digital products, such as mobile apps, websites, and e-commerce platforms. We will discuss the benefits of using AI in product design, development, and testing, and provide examples of companies that have leveraged AI to create innovative and successful products. While the benefits of AI are significant, we will also address the risks that businesses must consider as they implement AI. We will provide a link to resources that businesses can use to understand and mitigate these risks, including ethical considerations, data privacy, and security.

    NAVEEN JAIN: Privacy by Design: Building Privacy into Your Data Systems from the Ground Up

    As data scientists, it's our responsibility to build data systems that protect user privacy by design. In this talk, we'll explore the principles of "privacy by design" and discuss best practices for integrating privacy considerations into your data workflows from the ground up.

    We'll start by discussing key concepts such as anonymization and pseudonymization, and explore technical methods, tools, syntax, and data systems that data scientists can use to ensure that sensitive user information is protected throughout the data pipeline. We'll discuss techniques such as differential privacy and k-anonymity, and explore how to use tools such as Python's "pandas" library and SQL to implement privacy-enhancing features.

    By the end of this talk, attendees will have a clear understanding of how to implement privacy by design in their own data systems, from data collection and storage through analysis and reporting. They'll gain practical tips for anonymizing and pseudonymizing data, and leave with a toolkit of privacy-preserving techniques that they can use to build more responsible and ethical data systems.

    HYRUM ANDERSON, Robust Intelligence: Data and Model Supply Chain Risk in the AI Development Lifecycle

    With advancements in generative AI and other large language models, many organizations are forgoing in-house data collection and model development in favor of open source alternatives. While this makes AI accessible to more companies, it also opens the door to substantial unmanaged risk. In this talk, I review practical risks in poisoning webscale datasets, model repositories and models and review solutions for AI Supply Chain Risk Assurance.

    ERWIN QUIRING, International Computer Science Institute (ICSI) and Ruhr University Bochum: Dos and Don’ts of Machine Learning in Computer Security

    Machine learning algorithms enabled major breakthroughs in many different areas. This development has influenced computer security, spawning a series of work on learning-based security systems, such as for malware detection and vulnerability discovery. Despite great potential, machine learning in security is prone to subtle pitfalls that undermine its performance and render learning-based systems potentially unsuitable for security tasks and practical deployment.

    This talk gives an overview of this problem. Generic pitfalls in the design, implementation, and evaluation of learning-based security systems are presented. The talk discusses their prevalence in top-tier security research, and finally shows with case studies how individual pitfalls can lead to unrealistic performance and interpretations, obstructing the understanding of the security problem at hand.

  • Session Chair:

    Jennifer Zhu & Vidya Sagar Ravipati

    Confirmed Speakers:

    SEAN McCURDY, Pinterest: Intro to Pinterest Ads and ML Systems

    Pinterest is one of the world's largest visual discovery and bookmarking platforms, with over 450 million monthly active users and nearly 3b in yearly revenue. Behind this massive platform is a sophisticated machine learning (ML) system that helps to deliver relevant and personalized content to its users. In this talk, we will focus on Pinterest’s Ads Serving Funnel and explain how different components, such as targeting, retrieval, relevance, ranking, and marketplace work together to deliver relevant ads to Pinners. Finally, we will discuss how Pinterest leverages advancements in ML, most notably AutoML to unlock massive gains in revenue and Pinner ad engagement. Overall, attendees will gain a deeper understanding of how recommender systems are monetized and what considerations we made to balance outcomes for Pinners, Partners, and Pinterest.

    SAMAN SARRAF, Johnson & Johnson: Computer Vision Applications Across Industries: Wins and Challenges

    Numerous industries have shown interests in applying computer vision (CV) techniques to solve real-world use cases. Sports, media entertainment, autonomous driving, manufacturing, and healthcare are among top sectors that utilize such solutions for enhancing their audience engagement, security, and automation purposes. Successful computer vision end-to-end pipelines often include image classification, segmentation, object detection and tracking as well as pose estimation. However, using CV techniques usually face obstacles to meet business success criteria from both model performance and infrastructure perspective. In this presentation, we will review wins and challenges in this domain.

    RUIXUAN ZHANG, Airbnb: Machine Learning Driven User Growth and Lifecycle Engagement Marketing

    User growth has been a main focus for Airbnb as a two-sided marketplace. The ability of using inhouse targeting mechanisms to precisely identify the right users and proactively target them with relevant messages at the right time, largely enables us to improve marketing efficiency and drive user growth. In this talk, I will describe a machine learning system built at Airbnb that solves user growth problems at a large scale, which focuses on converting existing guests to hosts. In addition, I will also discuss how we used this ML framework in user lifecycle management and engagement marketing, to help increase campaign reach, user engagement and achieve high efficiency.

    APRIL LIU, Intuit: AI Application in Fighting Synthetic Account Fraud

    Fighting Synthetic Account Fraud with AI solution. Synthetic account fraud can have serious consequences for victims, such as negative credit score impact, financial loss and reputational harm. In recent years, a common ploy by fraudsters is to open up a large number of synthetic accounts that they can quickly recycle and use to cause damage (e.g., file fraudulent tax returns, send spam or phishing emails). At Intuit, we focus on developing AI-powered solutions for live risk assessment to effectively block fraudulent account creation at account sign-up time. In this talk, we’ll describe how we utilize short-term/long-term fraud models by leveraging AI auto-retrain capabilities to protect consumers and small businesses from harm.

  • Session Chair: Daniel Jerison
    Speakers:


    Moinak Bhaduri, Bentley University: Distribution-free, Online Change Detection in Multidimensional Point Processes Through Repeated Testing

    As we surface, probably momentarily, from the pandemic, other crises thwart normalcy: appalling inequality, climate calamity, distressed refugees, upped possibilities of a fresh Cold War. The enduring motif of our time is incessant chaos. Frequently, that chaos results when one type of a stationary system gives way to another. Change detection is mainly about estimating these points of deviation. In case a Poisson-type point process carries the system forward, I’ll offer a brand of online detection algorithms, engineered through permutations of trend switched statistics and a judicious application of false discovery rate control. Certain members of this family that remain asymptotically consistent and close to the ground truth (evidenced through some Hausdorff-similarity) are isolated to pinpoint estimated change locations. Efficient forecasting proves to be a natural corollary. Change point-based clustering tools will also be examined. I’ll describe how such analyses offer concrete definitions to vague objects like COVID “waves” and measure their enormity.


    WIRANTHE BANDARA HERATH, Drake University: Dimension Reduction for Vector Autoregressive Models

    The classical vector autoregressive (VAR) models have been widely used to model multivariate time series data, because of their flexibility and ease of use. However, the VAR model suffers from overparameterization particularly when the number of lags and the number of time series get large. There are several statistical methods of achieving dimension reduction of the parameter space in VAR models. In this talk, we introduce the reduced-rank VAR model (Velu et al., 1986; Reinsel and Velu, 2013) which restricts the rank of the parameter matrix in one direction, and the envelope VAR model (Wang and Ding, 2018) which is another solution to overcome the overparameterization problem. Then, we propose a new parsimonious VAR model by incorporating the idea of envelope models into the reduced-rank VAR. We show the strength and efficacy of the proposed model through some simulation studies and an economic data set.

    MUSTAFA HAJIJ, University of San Francisco: What is Topological Deep Learning?

    Over the past decade, deep learning has been remarkably successful at solving a massive set of problems on data types including images and sequential data. This success drove the extension of deep learning to other discrete domains such as sets, point clouds, graphs, 3D shapes, and discrete manifolds. While many of the extended schemes have successfully tackled notable challenges in each particular domain, the plethora of fragmented frameworks have created or resurfaced many long-standing problems in deep learning such as explainability, expressiveness and generalizability. Moreover, theoretical development proven over one discrete domain does not naturally apply to the other domains. Finally, the lack of a cohesive mathematical framework has created many ad hoc and inorganic implementations and ultimately limited the set of practitioners that can potentially benefit from deep learning technologies. This talk introduces the foundation of topological deep learning, a general mathematical framework for deep learning on topology spaces that expands the domains upon which deep learning protocols can be applied, all while maintaining intuitive conceptualization, implementation and relevance to a wide range of practical applications. We demonstrate the practical relevance of this framework with practical applications ranging from graph classification to mesh and segmentation.

    GREESHMA BALABHADRA, Stony Brook University: High-Frequency Risk Estimators Using Change Point Detection Methods

    We introduce new high-frequency volatility estimators that account for possible breakpoints in the spot volatility process. We further extend our method to the multivariate case of estimating covariance matrix in both high-frequency and high dimensional low-frequency space using the group fused LASSO method. In the univariate case, the resulting estimators are ℓ1-penalized versions of two classical high-frequency volatility estimators: quadratic variation and a jump-robust version of it, the bipower variation. We show that our estimators are consistent for the actual unobserved volatility. Numerically, the proposed estimators are fast in computations, and they accurately identify breakpoints close to the sample's end - both properties are desirable in the modern electronic trading environment. In terms of out-of-sample volatility prediction, qualitatively, the new estimators provide more smooth and realistic volatility forecasts; quantitatively, they outperform the aforementioned competitors and their extensions at various frequencies and forecasting horizons. In the multivariate case, the proposed estimator is used in minimum variance portfolio optimization and outperforms the benchmarks - realized covariance and multivariate heterogeneous autoregressive estimators in high-frequency and linear shrinkage and quadratic inverse shrinkage in low-frequency space.

  • Data Science as an Equalizer
    With the daily increase in data creation, improvements in technology, and the ever-growing need for informed decision-making, the growth of Data Science and Analytics is only expected to accelerate. As public and private organizations scramble to build their data teams, we focus on two timely issues in this panel discussion.

    One, data science, like most STEM fields, has dismaying diversity statistics. Does data science have an image problem? Is it considered impact-driven or vague and nerdy? Are there any unique strengths of the field that can improve diversity in the tech industry?

    Two, how can data science inform a more fair and equitable society? What are some recent developments in the public and private sectors that are using data to improve equity? Finally, what is hype vs. reality when we look toward the future of data science and its societal impact?

    Join us as we cover a broad range of opinions and perspectives on the role of data science as an equalizer.

    Session Chair
    Dr. Bushra Anum
    Director of Data Science & Analytics, Doximity

    Panelists:

    Elaine Zhou
    CTO, Change.org

    Carly Villareal
    Head of Data Science & Analytics Engineering, Nextdoor

    Megan Yahya, Senior Product Manager, Google Cloud

    Mark Freeman, Founder, On the Mark Data


    Disclaimer: The views and opinions expressed in this panel discussion are those of the speakers and do not necessarily reflect the views or positions of any entities they represent.

  • Join us for a reception where you can catch up with industry colleagues, make new friends, and create valuable connections. Music, refreshments, and hors d'oeuvres will be provided.


    Make sure to add this valuable networking experience to your conference schedule from 5:30 - 6:30 pm.

Tuesday, March 14

  • Registration is open from 8:00 AM - 4:00 PM each day.

  • The goal of precision medicine is to target the right treatments to the right patients at the right time. AI is the key technology required to unlock the potential of precision medicine.

    Most medical treatments are designed for the "average patient" as a one-size-fits-all-approach, which may be successful for some patients but not for others. Precision medicine, or personalized medicine, attempts to tailor disease prevention and treatment for a specific patient by considering individual genes, physiology, and environments. To accomplish this goal, large amounts of real-world clinical data must be analyzed.

    A learning healthcare system is a system in which data and knowledge generation is embedded into the routine healthcare delivery process, creating a continual stream of data that leads to improvement in personalized care. AI is a critical component of learning health systems that enable precision medicine.

    Session Chair:

    William Bosl

    Speakers:

    Vincent Liu, Kaiser Permanente Research: Augmented Intelligence: The Future of AI in Healthcare

    While there is tremendous excitement about the use of AI/ML in healthcare, there are still few studies that these tools drive improved patient outcomes. This talk will review Kaiser Permanente’s experience with deploying AI/ML predictive models in bedside care, including models that have saved hundreds of lives, and the implications of using this technology to enhance health and healthcare delivery.

    Sophia Wang, Stanford University: Envisioning the Future: Artificial Intelligence for Preserving Eyesight

    This presentation will discuss different ways in which artificial intelligence (AI) methodologies can be applied to ophthalmology healthcare data to predict and improve patient outcomes, with an emphasis on techniques for unstructured data modalities. Specific topics include the development of prediction models using structured and free-text data from electronic health records, to predict which glaucoma patients will progress and which patients with low vision have poor prognosis. Finally, we will present an AI system for automated analysis of cataract surgical videos to improve surgical training.

    Conrad Yiu, Trident.ai: AI in Medicine: What Are We Really Trying to Do?

    AI has the potential to improve the clinical care of patients; streamline operational processes to increase the efficiency, productivity, and quality of life of clinicians; and improve the economics for providers and payers which are all required in the US healthcare system. To date, AI has demonstrated limited ability to satisfy the needs in these areas. The fundamental question we discuss is - Can AI be applied to meet & achieve its promise?

  • Online controlled experiments (i.e., A/B tests) have seen a meteoric rise in prominence in the last decade. Generally speaking, such experiments seek to test and improve internet-based products and services using user-generated data to determine what works and what doesn’t. A/B tests have become an indispensable tool for major tech companies when it comes to maximizing revenue and optimizing the user experience, with some companies running hundreds of experiments engaging millions of users each day.

    In this online setting, and with a culture of testing as many ideas as possible, as quickly as possible, novel practical issues and modern challenges abound.

    This session brings together researchers and practitioners from both industry and academia to discuss such modern challenges and their work on them.

    Session Chair:

    Nathaniel Stevens

    Confirmed Speakers:

    NICK ROSS, University of Chicago: Hidden Integration Costs in Online Controlled Experimentation Platforms

    Data Science Practitioners have access to an array of different solutions when deciding how to implement an AB-testing platform for a product. At the top level of decision making is the traditional build vs. buy decision and, assuming the "buy" option there are a multitude of solution providers (Google Analytics, Split, Amplitude, etc.) which compete on both costs and features. A frequently overlooked aspect of this decision is that it is made against the backdrop of the analytics or data tracking system currently in place at an organization. Surprisingly, the success or failure of an AB-testing initiative is often determined by the integration between whatever AB-testing solution is chosen and the current analytics / tracking system. In this talk we will present a framework for analyzing this integration while will allow us to understand and predict the issues that may arise and, preemptively, avoid them.

    WENJING ZHENG, Netflix: Double Robust Causal Effect Generalization and Transportation under Covariate Shifts

    Being able to extrapolate the causal effect beyond the study population to a given target population allows us to increase the speed and utility of our learning. Building upon existing literature, we developed a framework and software tool for causal effect generalization and transportation. Generalizability addresses the problem of extrapolating the causal effect from the study population to a larger target population that contains it. Transportability addresses the problem of extrapolating the causal effect from a study population to a different target population. We describe some Netflix applications of this framework.

    STEVE HOWARD, The Voleon Group: Augmented Inverse Propensity Weighting for Randomized Experiments

    Augmented inverse propensity weighting (AIPW) is widely known in the context of observational studies, but is also a versatile and practical tool for variance reduction in randomized experiments. I'll present a unified view of AIPW estimation of the average treatment effect, relative lift, heterogeneous treatment effects, quantile treatment effects, and values of personalized policies using data from randomized experiments. Once one understands the "imputation principle" behind AIPW, there are "obvious" solutions to all of these problems, and those solutions tend to be statistically efficient and straightforward to implement with off-the-shelf algorithms.


    NATHANIEL STEVENS, University of Waterloo: General Additive Network Effect Models: A Framework for the Design and Analysis of Experiments on Networks

    As a means of continual improvement and innovation, online controlled experiments are widely used by internet and technology companies to test and evaluate product changes, new features, and to ensure that user feedback drives decisions. This is true of companies like Twitter, LinkedIn, and Facebook, large online social networks. However, experiments on networks are complicated by the fact that the stable unit treatment value assumption (SUTVA) no longer holds. Due to the interconnectivity of users in these networks, a user’s outcome may be influenced by their own treatment assignment as well as the treatment assignment of those they are socially connected with. The design and analysis of the experiment must account for this. We propose the General Additive Network Effect (GANE) model to jointly and flexibly model treatment and network effects. In this talk we discuss experimental design and analysis considerations in the context of the proposed model.

  • Session Chair: Daniel O’Connor

    Speakers:

    DILI EZEME, Ab-InBev: New Assortment Recommendation Approach for Differentiated Products

    Using the combination of transfer learning and random coefficients model, we develop a new methodology to overcome common practical challenges in recommendation systems such as cold start, sparsity, scalability, and lack of diversity in the data. On cold start, we use a machine learning approach to assign the products to clusters using the inherent product features. This clustering segments the products according to their competitiveness within and outside each cluster. This becomes a powerful insight that guides the econometric model to accurately simulate the behavior of the product in the presence of other products from within and outside the cluster. Traditionally, factorization machines (FM) solve the sparsity problem encountered in matrix factorization or collaborative filtering algorithms. However, FM suffers from dimensionality problems and scalability. To solve these sparsity and scalability challenges, we demonstrate how we leverage transfer learning by training using linear models and performing recommendations using non-linear complex models. This approach ensures we have a lean model of only 178 parameters running 1.3 billion data samples. Finally, we overcome the diversity challenge by leveraging the econometrics model which considers product, demographic, and account features. More importantly, our recommendation system considers product cannibalization and ensures that we generate optimal recommendations without hurting our portfolio's share or revenue. For validation, the proposed methodology outperformed FM in the same data in terms of MSE and RMSE.


    RUSHIL MANGLIK, University of San Francisco: Fine-Tuning Layout Parser deep learning model for Document Image Analysis and Table Extraction

    Layout Parser is a deep learning-based toolkit for document image analysis that lets us detect tables in documents which can potentially save significant labor time and costs associated with having to transcribe the tables manually. In this project, we fine-tune the layout parser deep learning model using the Stanford HPC system, Sherlock. The training data consists of a few labeled examples from recent EEO-1 reports that disclose the demographic workforce data including ethnicity by job category where a table or multiple tables in the filed EEO-1 reports have been manually labeled with a bounding box. We are interested in 1) detecting the tables with disclosed company information and 2) extracting information from the tables with Optical Character Recognition tools. In the future, we are interested in applying this pipeline to scanned images of historical documents and using NLP techniques on the extracted text.


    DIANE WOODBRIDGE, University of San Francisco: Bundle Recommender for Complimentary Menus

    Studies have shown that economies and their trends impact full-service restaurants [1]. This suggests that more people tend to stop using full-service restaurants and increase home-prepared meals during recessions.

    In order to improve user experience in cooking, diversify choices and maximize the basket size, the authors designed and developed a scalable data pipeline for recommending complimentary dishes and optimizing a shopping list based on a user’s initial entry.

    The developed workflow collects chef-curated menus from various web sources, develops natural language processing (NLP) models, and applies graph models to recommend the optimal complimentary menus. Furthermore, the authors optimized the quantities and prices to reduce the total cost and waste to enhance user experience [2].

    References
    [1] Lee, K. and Ha, I.S., 2012. Exploring the impacts of key economic indicators and economic recessions in the restaurant industry. Journal of Hospitality Marketing & Management, 21(3), pp.330-343.
    [2] Sethi, C., Vellera, M., Woodbridge, D.M.K. and Ahnn, J.J., 2022. Bundle Recommender from Recipes to Shopping Carts-Optimizing Ingredients, Kitchen Gadgets and their Quantities.
    Sethi C, Vellera M, Woodbridge DM, Ahnn JJ.


    SANGHAMITRA DEB, Chegg Inc.: Computer Vision Landscape at Chegg: Present and Future

    Millions of people all around the world Learn with Chegg. Education at Chegg is powered by the depth and diversity of the content that we have. A huge part of our content is in form of images. These images could be uploaded by students or could by content creators. Images contain text that is extracted using a transcription service. Very often uploaded images are noisy. This leads to irrelevant characters or words in the transcribed text. Using object detection techniques, we develop a service that extracts the relevant parts of the image and uses a transcription service to get clean text.

    In the first part of the presentation, I will talk about building an object detection model using YOLO for cropping and masking images to obtain a cleaner text from transcription. YOLO is a deep learning object detection and recognition modeling framework that is able to produce highly accurate results with low latency.

    In the next part of my presentation, I will talk about the building the Computer Vision landscape at Chegg. Starting from images on academic materials that are composed of elements such as text, equations, diagrams we create a pipeline for extracting image elements. Using state of the art deep learning techniques we create embeddings for these elements to enhance downstream machine learning models such as content quality and similarity.

  • Can We Reverse the Data Science Brain Drain?

    Data Science has been called the “The Sexiest Job of the 21st Century” and, along with Machine Learning Engineer, has stayed at the top of job reports for a decade. Three of the top 10 "most in-demand hard skills" for 2023 comprise data science (SQL, Python, and data analysis). This is amazing news for any data science student, and it's also a huge problem for the rest of us. These high paying job opportunities tend lure our best and brightest to redirect their creative data talents toward short-term, profit-maximizing corporate interests - like persuading people to buy things they don't need. In this talk, I hope to make at least a small dent in this 'big data brain drain' by highlighting the remarkable opportunities we have to contribute our data skills to solving some of the most existential challenges of our time.

  • Session Chair:

    Daniel Zielaski

    Confirmed Speakers:

    Chad Kimner, Meta Reality Labs: Navigating the Intersection of Personalization and Privacy: A Marketer's Perspective

    This session will take the audience on a light-hearted journey through the history of personalization in marketing over the last 20 years, with some personal stories along the way. Starting with the early days of direct mail marketing in the 1900s, we'll explore how personalization evolved with the rise of digital marketing and the use of cookies to track website visitors' behavior and interests. We'll then delve into the use of email personalization and social media advertising, before looking at the latest advances in artificial intelligence and machine learning.

    Building on this historical context, we'll then shift gears to focus on how marketers can operationalize customer-centric personalization by leveraging data science functions. Through a range of case studies, we'll explore topics such as personalization algorithms, measurement of personalized campaigns/programs, and experiments to improve and refine. Finally, we'll touch on how personalization may change in the era of increased data privacy regulation.

    Carson Forter, Twitch (Amazon): Putting Personalization to the Test: A Causal Inference Approach

    This presentation will focus on the use of causal inference experimental methods to test and refine personalization algorithms. We'll explore how data scientists can use techniques such as randomized controlled trials, quasi-experimental designs, and difference-in-differences analysis to determine causality and measure the impact of personalization strategies.

    By using these methods, data scientists can effectively isolate the effects of personalization from other variables and make data-driven decisions about which strategies are most effective. We'll also discuss how these techniques can be applied to other product and marketing data science projects to drive more informed decision-making.

    Sherry Guo, Salesforce: From 0 to 1: Leveraging Data Science Algorithms for Personalization

    This presentation will focus on how companies can leverage data science algorithms to move from a zero state to a state of effective personalization. We'll explore the use of the CatBoost machine learning model to develop personalized recommendations and experiences for customers. To achieve this, we'll discuss the importance of feature engineering, which involves selecting and extracting the right data points to feed into the model.

    But how can companies measure the effectiveness of their personalized campaigns? We'll also delve into personalization measurement, which involves tracking key metrics such as click-through rates, conversion rates, and customer engagement. By effectively measuring and analyzing these metrics, companies can fine-tune their personalization efforts and optimize for greater impact.
    Through a range of case studies and practical examples, we'll showcase the power of data science algorithms in driving effective personalization strategies. We'll also discuss the challenges and considerations that companies should keep in mind when implementing such solutions, including issues of data privacy and security. Overall, attendees can expect to gain valuable insights into the use of data science to drive personalized customer experiences.

    Maura Tuohy, Discord: The Fine Line: Balancing Personalization and Legal Compliance

    In the world of marketing, personalization is a powerful tool for engaging customers and driving sales. But as companies collect more data about their customers, they also run the risk of violating privacy laws and regulations. In this talk, we'll explore the fine line between personalization and legal compliance, examining high-profile examples of companies that have crossed the line and suffered the consequences. From Facebook's Cambridge Analytica scandal to Target's data breach, we'll discuss the legal and ethical considerations that companies must navigate when implementing personalization strategies. We'll also provide practical tips for staying on the right side of the law, including best practices for data collection, consent, and transparency. By the end of this talk, attendees will have a clear understanding of how to balance personalization and legal compliance, and be better equipped to build responsible and ethical marketing strategies.

  • Robo advisors, which utilize personal data and AI to customize wealth management products, have arisen as the next generation of investment and wealth management services. This session shall explore the cutting edge of how reinforcement learning is being used to advance the application of financial models and AI in the wealth management industry.

    Session Chair:

    Matthew Dixon, Illinois Institute of Technology

    Confirmed Speakers:

    Sanjiv Das, Santa Clara University

    Karén Chaltikian, MoneyLion

    Sanjiv Das: Optimizing The Probabilities Of Obtaining Single And Multiple Financial Goals
    We develop a dynamic investment strategy to optimize the probability of attaining a single financial goal like staying solvent in retirement by using either dynamic programming or reinforcement learning. This is extended to optimally attaining given probabilities for multiple competing financial goals. This multiple goals case entails not only the optimal dynamic investment strategy but also optimally choosing when to obtain a financial goal versus when to forgo the goal to improve the chances of obtaining more important later goals. A byproduct of this algorithm is a new artifact, the efficient goal probability frontier, which complements the well-known mean-variance efficient frontier.


    Karen Chaltikian: Wealth Management Benchmarks
    Investment management performance is easily measured at any frequency and easily
    compared to publicly traded benchmarks such as S&P500 and Barclays Agg indices. In contrast, wealth management is concerned with custom portfolios evolving over multi-decade horizons. Such long horizons make it impractical to assess the quality of employed strategies ex-post, while singularly personalized nature of the objective - funding a set of goals specific to each individual - makes it impossible to create a single observable benchmark similar to an index. We suggest that indexing should be done not at the level of portfolios but strategies, in a way reminiscent of target date funds, and evaluation should be done using Monte-Carlo simulations. We propose a simple methodology for building personalized benchmarks based on a combination of user starting financial state and arbitrary list of goals. Such benchmarks can be used to realistically evaluate a client’s current financial situation and serve as a point of departure for providing nuanced advice.

    Matthew Dixon: Time Consistent Risk-Aware Q-Learning of Optimal Consumption under Epstein-Zin Preferences
    We present a class of reinforcement learning algorithms for optimal consumption under intertemporal substitution and risk adversity preferences. The classical setting of Epstein Zin utility preferences is cast into a dynamic risk measure framework and shown to give dynamic risk-aware performance criteria which are time consistent. This permits the robust approximation of the optimal consumption problem as a discrete time Markov Decision Process. We present a risk-sensitive Q-Learning algorithm suitable for non-linear monotone risk measures and benchmark its policy estimation convergence properties on an optimal wealth consumption problem against Least Squares Monte-Carlo and binomial tree methods.

  • Session Chair: Shan Wang
    Speakers:


    ZIRUI ZHANG, University of California, Irvine: Parameter Inference in Diffusion-Reaction Models of Glioblastoma Using Physics-Informed Neural Networks

    Glioblastoma is an aggressive brain tumor that proliferates and infiltrates into the surrounding normal brain tissue. The growth of Glioblastoma is commonly modeled mathematically by diffusion-reaction type partial differential equations (PDEs). These models can be used to predict tumor progression and guide treatment decisions for individual patients. However, this requires parameters and brain anatomies that are patient specific. Inferring patient specific biophysical parameters from medical scans is a very challenging inverse modeling problem because of the lack of temporal data, the complexity of the brain geometry and the need to perform the inference rapidly in order to limit the time between imaging and diagnosis. Physics-informed neural networks (PINNs) have emerged as a new method to solve PDE parameter inference problems efficiently. PINNs embed both the data the PDE into the loss function of the neural networks by automatic differentiation, thus seamlessly integrating the data and the PDE. In this work, we use PINNs to solve the diffusion-reaction PDE model of glioblastoma and infer biophysical parameters from numerical data. The complex brain geometry is handled by the diffuse domain method. We demonstrate the efficiency, accuracy and robustness of our approach.


    SOURADIPTO GHOSH DASTIDAR, University of Minnesota Twin Cities: Correcting for Spatial Correlation in Functional Connectivity from Task-based fMRI

    Task-based fMRI, that measures BOLD signal changes between control and task-stimulated states, can be used to locate specific brain locations in the different regions of interest (ROIs) associated with different neurodevelopmental disorders. Due to immense computational complications, most modern neuroimaging data analysis use what is called a massive univariate regression modelling for each labelled location in the ROIs. Recent developments have been trying to incorporate spatial dependencies between different brain regions and implementing post-hoc inference techniques to reduce false positive errors. However, such methods involve predefining neighborhoods, which might not be explicitly clear for functional network pairs. In order to correct for spatial autocorrelation among functional connectivity pairs, we propose a measure of functional overlap and invert it to obtain a functional distance. Finally, we model the brain-wise spatial autocorrelations, compute the multivariate test-statistics, use the functional distance to compute clusterwise test-statistics and implement non-parametric resampling approach to reduce type I error rates and improve reproducibility.

    The primary data comes from a study of fMRI images on macaques, where we have information about the functional connectivity on each of the 3321 ROI pairs coming from 82 ROIs. The response variable is lutein- a xanthophyll carotenoid found in the brain tissue. To test the effectiveness of the methodology we plan use real data from the ABCD study (N= 11,877), a longitudinal study of brain development, and an independent ADHD-enriched sample (7-11 yo, N= 656/424 total/ADHD) and test the model predicting scores of ADHD-like behaviors.


    VIGNESH RAVINDRANATH, University of California, San Francisco: Optimizing Treatment Selection in Crohn’s Disease Using Patient-Specific Features: An Individual Participant Data Meta-Analysis of Fifteen Randomized Controlled Trials

    Crohn’s disease is characterized by diverse manifestations that reflect intrinsic differences in disease biology and treatment responsiveness. Prior meta-analyses have found anti-tumor necrosis factor (TNF) drugs to be more efficacious than other drug classes. However, these findings are based on cohort-averaged effects and have ignored the role of patient-level variation in determining treatment outcomes. We sought to model personalized treatment options for patients with Crohn’s disease and characterize patient subgroups using participant-level data from randomized trials. Following a search of clnicaltrials.gov we obtained access to participant-level data from 15 trials (N=5703) involving adults with moderate-to-severe Crohn’s disease treated with a biologic (anti-TNFs, anti-interleukin-12/23s, or anti-integrins). We used sequential regression and simulation to separately model the placebo- and drug-attributable effects as a function of drug class, demographics, and disease-related features. We used these models to simulate patient-level outcomes following treatment with different drug classes. We performed hypothesis tests to compare these potential outcomes and to quantify the evidence favoring personalized treatment selection over several null models, 1) that each drug class would have the same efficacy irrespective of patient-level variation (one-size-fits-all), and 2) that pairs of drug classes would have the same efficacy in each individual patient. We used a p-value threshold of 0.05 to categorize patients into subgroups defined by treatment preferences. We performed queries on the University of California Health Data Warehouse to measure the prevalence of these subgroups and quantify the potential real-world impact of these findings. Lastly, we prototyped a decision support tool that uses manual inputs and OMOP-formatted data to recommend treatments for individual patients.

  • Session Chair: Dr. Robert Clements

    2022 was arguably the year of generative AI. Text-to-image models such as DALL-E 2, Midjourney and Stable Diffusion, trained on massive amounts of data available on the internet, have made it easy and fun for anybody to generate an image simply by writing out a description of what they want to see. Join us as we discuss the ethical issues of AI-generated art, such as the inclusion of other people's images in the training of these models, the uses of AI-generated art in the public sphere, and the impact on the creative jobs landscape.

    Speakers:

    Quinn Keck, Disney Streaming

    Paul Kim, UC Santa Barbara

    Chris Brooks, University of San Francisco

    David Aughenbaugh, Factory VFX and EyeTripImages.com

  • Join us for a reception where you can catch up with industry colleagues, make new friends, and create valuable connections. Music, refreshments, and hors d'oeuvres will be provided.

    Make sure to add this valuable networking experience to your conference schedule from 5:30 - 6:30 pm.