Asked by john

1. How does machine learning work?



2. Think of your favorite app. How could interaction bias affect how the app works?



3. Latent bias is based on past events or data. How do you think this could be changed or avoided?



4. How can selection bias be limited?

All Answers 1

Answered by GPT-5 mini AI
1) How does machine learning work?
- At a high level, machine learning (ML) builds models that make predictions or decisions by learning patterns from data instead of being explicitly programmed.
- Typical steps:
1. Collect and prepare data (cleaning, feature extraction/selection).
2. Choose a model type (e.g., linear regression, decision trees, neural networks).
3. Define an objective (loss) that measures how wrong the model is.
4. Train the model by adjusting parameters to minimize that loss (optimization).
5. Evaluate on held-out data to check generalization and tune hyperparameters.
6. Deploy the model and monitor performance in production, retraining as needed.
- Main paradigms: supervised learning (learn from labeled examples), unsupervised learning (find structure in unlabeled data), reinforcement learning (learn by trial and feedback).
- Important concepts: overfitting (model memorizes training data), underfitting (model too simple), bias–variance tradeoff, and the need for representative, high-quality data.

2) Interaction bias and an example app
- Interaction bias occurs when users’ interactions shape future system behavior in a way that amplifies certain signals or narrows the system’s outputs.
- Example: TikTok or YouTube recommendations. If a user clicks short, sensational videos, the recommender shows more of the same. Those interactions train the system further, reinforcing a narrow type of content (echo chamber), possibly reducing diversity and surfacing low-quality or extreme content.
- Consequences: feedback loops, filter bubbles, content polarization, and neglect of minority interests.
- Mitigations: introduce exploration (occasionally show diverse or randomly selected items), apply diversity and novelty constraints, use decay on old interaction weights, incorporate explicit user controls (like “show me more of X” vs “less of Y”), and monitor long-term engagement quality rather than only short-term clicks.

3) Changing or avoiding latent bias (bias from past data)
- Latent bias arises when historical data reflects past discrimination, social inequities, or skewed practices, which models then reproduce.
- Ways to reduce it:
- Audit data and models to detect biased outcomes across sensitive groups (e.g., race, gender, age).
- Improve data collection to be more representative (targeted sampling, filling gaps).
- Remove or mask explicit and proxy features tied to protected attributes where appropriate.
- Reweight or resample training data to compensate for underrepresented groups.
- Use fairness-aware algorithms and constraints (e.g., equalized odds, demographic parity) or adversarial debiasing.
- Apply causal analysis to separate spurious correlations from true causal relationships.
- Include human oversight and stakeholder input to surface historical harms and context.
- Continuously monitor and update models as society and data change.
- Note: fully "erasing" latent bias is hard — the pragmatic goal is to identify, mitigate, and document risks.

4) Limiting selection bias
- Selection bias happens when the sample used to train or evaluate a model is not representative of the target population.
- Ways to limit it:
- Design sampling strategies that aim for randomness or proper stratification so all subgroups are represented.
- Use stratified or quota sampling to ensure minority groups aren’t under-sampled.
- In experiments, use randomized assignment (A/B testing) and minimize differential nonresponse.
- If the available sample is biased, apply statistical adjustments: weighting, post-stratification, inverse probability weighting, or propensity-score methods to correct for known selection mechanisms.
- Combine multiple data sources and validate results on independent datasets.
- Document inclusion/exclusion criteria and run sensitivity analyses to understand how selection choices affect outcomes.
- Actively recruit underrepresented populations or oversample them and then reweight during analysis.
- Regularly check model performance across subgroups to detect residual selection effects.

If you want, tell me an app you use and I’ll sketch concrete interaction-bias risks and mitigation steps for that specific case.