m. a.
Details
Skills
Large Language Model (LLM) Evaluation & Benchmarking
Prompt Engineering & Prompt Quality Analysis
Multilingual Linguistic Analysis (Arabic–English)
AI Output Review (Accuracy, Relevance, Bias, Safety)
Data Annotation, Labeling & Quality Assurance
Speech Data QA (Transcription Review, Speaker Segmentation)
Comparative Model Evaluation (Search vs Conversational AI)
Cultural & Linguistic Alignment Assessment
Instruction Following & Guideline Compliance
Error Analysis & Structured Feedback Reporting
Remote Collaboration & Task-Based Productivity
Attention to Detail & High-Volume Review Accuracy
About
I have hands-on experience in AI data operations, linguistic analysis, and LLM evaluation across multiple global platforms. I worked remotely with Innodata Inc. as a Linguistics Expert (Arabic), where I optimized large language model performance through prompt engineering, linguistic annotation, multilingual data validation, and bias detection, while evaluating outputs for factual accuracy, cultural sensitivity, and linguistic integrity.
Additionally, I contributed to several AI training and evaluation projects with OneForma, Upwork Enterprise, and Invisible Technologies, including Arabic voice data QA, transcription correction, speaker segmentation, and real-time comparative evaluations of conversational AI models (Search Live vs. Gemini Live). My responsibilities included assessing reasoning quality, relevance, safety, and instruction adherence, as well as documenting failure cases and improvement insights according to strict QA guidelines.
Alongside my AI work, I have experience in instructional design and technical education as a remote instructor, delivering Arabic-language courses in web design and computer science. I am currently pursuing a B.Sc. in Biomedical Sciences with a focus on Computational Biology, which strengthens my analytical thinking, data handling, and interdisciplinary approach to AI and technology-driven projects.