Niharmahesh commited on
Commit
5d16948
·
verified ·
1 Parent(s): 1285b60

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +11 -9
app.py CHANGED
@@ -64,15 +64,17 @@ def display_work_experience():
64
  st.markdown('## Work Experience')
65
  st.write("""
66
 
67
- **Turing, San Jose, CA, USA**
68
- March 2025 - Present
69
- - **Data Scientist**
70
- - Successfully shipped a custom evaluation-benchmark dataset for Gemini 3.0, aggregating data sources across multiple
71
- domains (mathematics, finance, chemistry, biology) spanning educational levels from high-school through PhD
72
- - Collaborate with AI engineers, product teams, and academic researchers to align research initiatives with business objectives and deliver data-driven solutions across cross-functional teams
73
- - Fine-tuned Qwen model on multi-GPU cluster using advanced architecture, boosting overall performance by 12% and achieving significant accuracy improvements across 15+ analytical categories
74
- - Design and implement robust evaluation pipelines incorporating quality assessments, performance benchmarking, and bias mitigation techniques to enhance model reliability and fairness
75
-
 
 
76
  **San Jose State University, San Jose, CA, USA**
77
  August 2024 - December 2024
78
  - **Teaching Assistant**
 
64
  st.markdown('## Work Experience')
65
  st.write("""
66
 
67
+ **Turing, San Jose, CA, USA**
68
+ March 2024 - Present
69
+ - **Data Scientist & Applied AI Engineer**
70
+ - Collaborated with AI engineers, product teams, researchers, and Google DeepMind team to integrate LLM evaluation systems into production workflows using PyTorch and distributed computing
71
+ - Engineered comprehensive evaluation benchmarks for Gemini 3.0 by analyzing reasoning loss patterns and image loss patterns in state-of-the-art Vision-Language Models (VLMs) including o3 and Gemini 2.5 Pro, developing custom datasets across multiple domains (mathematics, finance, chemistry, biology) spanning educational levels from high-school through PhD with statistical validation methods
72
+ - Implemented advanced LLM fine-tuning strategies for Qwen model including Parameter-Efficient Fine-Tuning (PEFT) with LoRA and 2-stage whole model training on multi-GPU clusters, achieving 12% performance improvement across 15+ categories
73
+ - Developed "auto hinter" system to improve LLM reasoning, guiding models towards correct answers based on question complexity, resulting in 8% performance increment on PhD-level questions
74
+ - Built "auto rater" system to assess responses from leading models like Gemini 2.5 Pro and o3 custom builds, scoring across four key dimensions: completeness, coherence, clarity, and correctness
75
+ - Applied advanced model compression techniques including quantization and distillation methods to optimize inference performance while maintaining model accuracy for production-ready LLM deployment
76
+ - Designed robust evaluation pipelines incorporating ROC curve analysis, performance benchmarking, bias mitigation, and RMSE validation to ensure model reliability and efficiency
77
+
78
  **San Jose State University, San Jose, CA, USA**
79
  August 2024 - December 2024
80
  - **Teaching Assistant**