SonicJobs Logo
Left arrow iconBack to search

Senior Research Data Scientist, Core Safety Data Science

Google
Posted 13 hours ago, valid for 3 days
Location

San Francisco, CA 94102, US

Salary

$174,000 - $252,000 per year

Contract type

Full Time

By applying, a Sonicjobs account will be created for you. Sonicjobs's Privacy Policy and Terms & Conditions will apply.

SonicJobs' Terms & Conditions and Privacy Policy also apply.

Sonic Summary

info
  • The position requires a Master's degree in a quantitative field and a minimum of 5 years of relevant experience, or 3 years with a PhD.
  • Candidates should have 4 years of experience in data analysis as a statistician or data scientist, along with proficiency in statistical software and database languages.
  • The role focuses on improving user protections against online abuse through data science capabilities and collaboration with product teams.
  • The US base salary for this full-time position ranges from $174,000 to $252,000, excluding bonus, equity, and benefits.
  • This position also emphasizes the development of methodologies for Generative AI safety and the evaluation of content safety classifiers.

Minimum qualifications:

  • Master's degree in Statistics, Data Science, Mathematics, Physics, Economics, Operations Research, Engineering, or a related quantitative field, or equivalent practical experience.
  • 5 years of experience using analytics to solve product or business problems, coding (e.g., Python, R, SQL), querying databases, or statistical analysis, or 3 years of experience with a PhD degree.
  • 4 years of experience in data analysis or related fields as a statistician or data scientist.
  • Experience with statistical software (e.g., R, Python, MATLAB, pandas) and database languages (e.g., SQL).
  • Experience with statistical methodologies.

Preferred qualifications:

  • PhD degree in Statistics, Data Science, Mathematics, Physics, Economics, Operations Research, Engineering, or a related quantitative field.
  • 8 years of work experience using analytics to solve product or business problems, coding (e.g., Python, R, SQL), querying databases or statistical analysis, or 6 years of work experience with a PhD degree.
  • Experience in training, validating or optimizing language models or LLM-based classifiers.
  • Experience analyzing multi-modal data (image, audio, or video).
  • Experience of GenAI safety and red-teaming.

About the job:

User Protection is an organization dedicated to protecting Google's users from abuse, account compromise and other harms online. Our team works with the Content Safety (CS) and User Protection Platform and Services (UPS) which develops tools to protect users from abusive content at scale, often leveraging AI technology to do so. Our team provides data science capabilities to these two organizations, and works directly with product and engineering to evaluate, understand, and improve the quality of our protections. Organizationally, we are a part of a large data science team in Core, which provides ample opportunities for knowledge sharing, development, and learning from other data scientists working in adjacent domains.

CS and UPS equip Google products with tools to protect users from abuse and harm. As a Data Scientist working with CS and UPS, you'll be helping to evaluate, understand, and improve our abuse protections - which are generally built with and for AI tools. We work closely with cross-functional product teams on specific content safety classifiers, but also on generic strategies and tooling for understanding content safety classifiers.

Our team is designing safety data evaluations and safety mitigation evaluations, including LLM-as-judge, prompt injection, and Responsible AI testing. We also work with flagship GenAI product teams on understanding Google-wide GenAI safety postures in production traffic.The US base salary range for this full-time position is $174,000-$252,000 + bonus + equity + benefits. Our salary ranges are determined by role, level, and location. Within the range, individual pay is determined by work location and additional factors, including job-related skills, experience, and relevant education or training. Your recruiter can share more about the specific salary range for your preferred location during the hiring process.

Please note that the compensation details listed in US role postings reflect the base salary only, and do not include bonus, equity, or benefits. Learn more about benefits at Google.

Responsibilities:

  • Solve ambiguous problems in the Generative Artificial Intelligence safety space, including agent-based safety.
  • Develop quantitative methodologies to curate training data and evaluation data from synthetic data and real-world production data for improving content safety mitigations. Design and evaluate models to mathematically express and solve defined problems with limited precedent.
  • Drive cross-functional alignment on measuring violation rates and unjustified refusals across multiple flagship Generative Artificial Intelligence product surfaces. Identify and clarify business or product questions.
  • Provide feedback and refine business questions into tractable analysis, evaluation metrics, or mathematical models. Drive clarity and coherence in understanding safety at scale across Google.
  • Own the process of gathering, extracting, and compiling data across sources (e.g., SQL, R, Python). Format, re-structure, or validate data to ensure quality, and review the dataset to ensure it is ready for analysis.



Learn more about this Employer on their Career Site

Apply now in a few quick clicks

By applying, a Sonicjobs account will be created for you. Sonicjobs's Privacy Policy and Terms & Conditions will apply.

SonicJobs' Terms & Conditions and Privacy Policy also apply.