<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: Servin Osmanov</title>
    <description>The latest articles on Forem by Servin Osmanov (@servin_osmanov).</description>
    <link>https://forem.com/servin_osmanov</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F3583111%2Fb3feb8b3-8b2e-4b44-96f8-41d587423dbf.jpg</url>
      <title>Forem: Servin Osmanov</title>
      <link>https://forem.com/servin_osmanov</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/servin_osmanov"/>
    <language>en</language>
    <item>
      <title>My Journey Improving a TTS Model for the Crimean Tatar Language</title>
      <dc:creator>Servin Osmanov</dc:creator>
      <pubDate>Fri, 07 Nov 2025 19:42:08 +0000</pubDate>
      <link>https://forem.com/servin_osmanov/my-journey-improving-a-tts-model-for-the-crimean-tatar-language-53f7</link>
      <guid>https://forem.com/servin_osmanov/my-journey-improving-a-tts-model-for-the-crimean-tatar-language-53f7</guid>
      <description>&lt;p&gt;When you work with machine learning, success often hides behind hours of frustration, countless errors, and broken pipelines. This project — improving the &lt;strong&gt;Crimean Tatar TTS (Text-to-Speech)&lt;/strong&gt; model — was exactly that kind of journey. What started as a small experiment to fine-tune an existing model turned into a full-scale debugging adventure that taught me more about data integrity, audio processing, and patience than any tutorial could.&lt;/p&gt;

&lt;p&gt;In my previous article, &lt;a href="https://dev.to/servin_osmanov/why-language-tech-matters-developing-ai-tools-for-small-languages-583h"&gt;Why Language Tech Matters: Developing AI Tools for Small Languages&lt;/a&gt;, I explored how AI can empower low-resource languages. This piece continues that journey with a hands-on look at improving TTS models.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu4vpl5k2r345s9eueqho.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fu4vpl5k2r345s9eueqho.png" alt="Sevil tts model improvement" width="800" height="533"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  The Starting Point: A Model That Worked — but Only Partially
&lt;/h2&gt;

&lt;p&gt;My goal was simple: improve the voice model “&lt;strong&gt;Sevil&lt;/strong&gt;” for the Crimean Tatar language. I had already worked with similar voices — “&lt;strong&gt;Arslan&lt;/strong&gt;” and “&lt;strong&gt;Abibullah&lt;/strong&gt;” — using Hugging Face datasets like &lt;code&gt;speech-uk/tts-crh-arslan&lt;/code&gt; and &lt;code&gt;speech-uk/tts-crh-abibullah&lt;/code&gt;.&lt;/p&gt;

&lt;p&gt;The first training attempts went well — &lt;strong&gt;loss around 0.283&lt;/strong&gt;, results acceptable. But something didn’t add up. The dataset had &lt;strong&gt;1,566 audio files&lt;/strong&gt;, yet the training logs showed only &lt;strong&gt;415 were being used&lt;/strong&gt; — about &lt;strong&gt;26.5%&lt;/strong&gt; of the total.&lt;br&gt;&lt;br&gt;
That meant &lt;strong&gt;almost three-quarters of my data was silently ignored&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;At first, I thought it was a fluke. Then I realized it was a &lt;strong&gt;systemic problem&lt;/strong&gt; in the Hugging Face &lt;code&gt;datasets&lt;/code&gt; API when loading compressed audio from Parquet files.  &lt;/p&gt;
&lt;h2&gt;
  
  
  Diagnosing the Problem
&lt;/h2&gt;

&lt;p&gt;When I loaded the dataset through &lt;code&gt;datasets.load_dataset()&lt;/code&gt;, most files threw errors:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;Error: "Error while decoding audio"
Error: "Audio file appears to be empty"
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;That didn’t make sense — the audio bytes were clearly present in the Parquet files.&lt;br&gt;&lt;br&gt;
After checking manually with &lt;code&gt;pandas.read_parquet()&lt;/code&gt;, I confirmed the data was there. The problem wasn’t in the files — it was in how the decoder handled them.&lt;/p&gt;

&lt;p&gt;That’s when I realized: the &lt;strong&gt;datasets API couldn’t decode raw audio bytes correctly&lt;/strong&gt;. The data was fine, but the pipeline was broken.&lt;/p&gt;
&lt;h2&gt;
  
  
  Turning Bytes into Sound
&lt;/h2&gt;

&lt;p&gt;At this point, I tried everything.&lt;br&gt;&lt;br&gt;
I extracted the bytes manually, saved them as &lt;code&gt;.raw&lt;/code&gt; files, and tried to load them with &lt;code&gt;librosa&lt;/code&gt; and &lt;code&gt;soundfile&lt;/code&gt;. Nothing worked.&lt;br&gt;&lt;br&gt;
Without proper &lt;strong&gt;metadata&lt;/strong&gt; (sample rate, channels, encoding), the files were unreadable.&lt;/p&gt;

&lt;p&gt;Eventually, I discovered the solution: &lt;strong&gt;FFmpeg&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;By using known dataset parameters —&lt;br&gt;&lt;br&gt;
&lt;code&gt;sample rate: 16000 Hz&lt;/code&gt;, &lt;code&gt;channels: mono&lt;/code&gt;, &lt;code&gt;format: PCM 16-bit little-endian&lt;/code&gt; —&lt;br&gt;&lt;br&gt;
I could convert all &lt;code&gt;.raw&lt;/code&gt; files into clean &lt;code&gt;.wav&lt;/code&gt; audio.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight shell"&gt;&lt;code&gt;ffmpeg &lt;span class="nt"&gt;-f&lt;/span&gt; s16le &lt;span class="nt"&gt;-ar&lt;/span&gt; 16000 &lt;span class="nt"&gt;-ac&lt;/span&gt; 1 &lt;span class="se"&gt;\&lt;/span&gt;
       &lt;span class="nt"&gt;-i&lt;/span&gt; sevil_0000.raw &lt;span class="se"&gt;\&lt;/span&gt;
       sevil_0000.wav
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;And just like that — &lt;strong&gt;1,566 files successfully converted&lt;/strong&gt;.&lt;br&gt;&lt;br&gt;
No corruption. No decoding errors. 100% validation success.&lt;/p&gt;

&lt;p&gt;It was the moment of breakthrough — the kind that makes you sit back, smile, and realize you’ve just solved a problem that haunted you for two days straight.&lt;/p&gt;
&lt;h2&gt;
  
  
  Training the Model — Again
&lt;/h2&gt;

&lt;p&gt;With clean audio finally ready, I retrained the Sevil model from scratch, this time using &lt;strong&gt;all 1,566 recordings&lt;/strong&gt;.&lt;/p&gt;

&lt;p&gt;Training setup (based on my previous configs for “Arslan”):&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;num_train_epochs = 500
batch_size = 4
learning_rate = 1e-4
fp16 = True
warmup_steps = 2000
save_steps = 2000
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;The progress was promising:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Loss dropped from &lt;strong&gt;1.14 → 0.80 → 0.50 → 0.27&lt;/strong&gt;
&lt;/li&gt;
&lt;li&gt;The voice quality improved with every epoch
&lt;/li&gt;
&lt;li&gt;And then… it crashed at &lt;strong&gt;78%&lt;/strong&gt; completion.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The culprit? A familiar one for SpeechT5 users —&lt;br&gt;&lt;br&gt;
&lt;code&gt;RuntimeError: torch.cat(): expected a non-empty list of Tensors&lt;/code&gt;&lt;br&gt;&lt;br&gt;
It turned out to be a bug in &lt;strong&gt;guided_attention_loss&lt;/strong&gt;, a component that sometimes fails with uneven sequence lengths.&lt;/p&gt;
&lt;h2&gt;
  
  
  Fixing the Crash
&lt;/h2&gt;

&lt;p&gt;Instead of starting over, I resumed training from the last checkpoint (step 16,000) and simply &lt;strong&gt;disabled guided attention&lt;/strong&gt;:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;model.config.use_guided_attention_loss = False
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;That one line saved the project.&lt;br&gt;&lt;br&gt;
Training resumed, completed &lt;strong&gt;98% of the full cycle&lt;/strong&gt;, and stabilized with a final &lt;strong&gt;loss of 0.267&lt;/strong&gt; — a small numerical improvement, but a big qualitative one.&lt;br&gt;&lt;br&gt;
The model became more consistent and robust across new data.&lt;/p&gt;

&lt;h2&gt;
  
  
  Comparing Results
&lt;/h2&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Model&lt;/th&gt;
&lt;th&gt;Data Used&lt;/th&gt;
&lt;th&gt;Loss&lt;/th&gt;
&lt;th&gt;Success&lt;/th&gt;
&lt;th&gt;Notes&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Sevil v1&lt;/td&gt;
&lt;td&gt;415 files (26.5%)&lt;/td&gt;
&lt;td&gt;0.276&lt;/td&gt;
&lt;td&gt;47% test&lt;/td&gt;
&lt;td&gt;Trained on partial data&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Sevil v2&lt;/td&gt;
&lt;td&gt;1,566 files (100%)&lt;/td&gt;
&lt;td&gt;0.267&lt;/td&gt;
&lt;td&gt;100% test&lt;/td&gt;
&lt;td&gt;Fully trained, stable&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;The difference wasn’t just in metrics — it was in &lt;strong&gt;confidence&lt;/strong&gt;.&lt;br&gt;&lt;br&gt;
Sevil v2 generalized better, produced smoother intonation, and maintained pronunciation consistency even on unseen words.&lt;/p&gt;

&lt;h2&gt;
  
  
  Lessons Learned
&lt;/h2&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Never trust good metrics without checking data coverage.&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
My “good” baseline was trained on just 26% of the data.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;FFmpeg is a lifesaver.&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
It solved what specialized libraries couldn’t.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Validate every single file.&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
Automation saved hours of manual checking.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Guided attention loss is optional.&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
Sometimes stability matters more than theoretical accuracy.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Document everything.&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
By keeping track of every attempt — successful or not — I could understand the full story, not just the happy ending.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  Why This Project Matters
&lt;/h2&gt;

&lt;p&gt;For me, this wasn’t just about fixing one dataset. It was about enabling a &lt;strong&gt;low-resource language&lt;/strong&gt; — Crimean Tatar — to have a better voice in the digital world.&lt;br&gt;&lt;br&gt;
Improving the Sevil model means clearer pronunciation, smoother prosody, and better accessibility for learners and native speakers alike.&lt;/p&gt;

&lt;p&gt;And for anyone working with custom TTS datasets:&lt;br&gt;&lt;br&gt;
Check your files, validate your data, and don’t give up when your model crashes at 78%.&lt;br&gt;&lt;br&gt;
That crash might be the best teacher you’ll ever have.&lt;/p&gt;




&lt;p&gt;&lt;strong&gt;Author:&lt;/strong&gt; &lt;em&gt;Servin Osmanov&lt;/em&gt;&lt;br&gt;&lt;br&gt;
Lead Fullstack Python / ReactJS Engineer&lt;br&gt;&lt;br&gt;
AI researcher and TTS developer for low-resource languages&lt;br&gt;&lt;br&gt;
Project: &lt;code&gt;servinosmanov/tts-crh-sevil-fixed&lt;/code&gt; on Hugging Face&lt;/p&gt;

</description>
      <category>ai</category>
      <category>huggingface</category>
      <category>programming</category>
      <category>nlp</category>
    </item>
    <item>
      <title>My Journey as a Judge at CBIT Hacktoberfest 2025 — Lessons from the Other Side of the Table</title>
      <dc:creator>Servin Osmanov</dc:creator>
      <pubDate>Thu, 30 Oct 2025 18:29:37 +0000</pubDate>
      <link>https://forem.com/servin_osmanov/my-journey-as-a-judge-at-cbit-hacktoberfest-2025-lessons-from-the-other-side-of-the-table-78e</link>
      <guid>https://forem.com/servin_osmanov/my-journey-as-a-judge-at-cbit-hacktoberfest-2025-lessons-from-the-other-side-of-the-table-78e</guid>
      <description>&lt;p&gt;When I first started attending hackathons, I was always the one coding, building, and pitching. This year, for the first time, I got to experience the other side — as a &lt;strong&gt;judge&lt;/strong&gt; at the &lt;a href="https://cbit-hacktoberfest25.devpost.com/?ref_feature=challenge&amp;amp;ref_medium=discover#prizes" rel="noopener noreferrer"&gt;CBIT Hacktoberfest Hackathon 2025&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;It was an incredible 24-hour online event organized by the &lt;strong&gt;CBIT Open Source Community&lt;/strong&gt;, celebrating open-source culture and collaboration as part of the global &lt;strong&gt;Hacktoberfest&lt;/strong&gt;. The event gathered hundreds of students from universities across the world — developers, designers, and dreamers ready to learn, code, and share.&lt;/p&gt;

&lt;h2&gt;
  
  
  What the Event Was About
&lt;/h2&gt;

&lt;p&gt;Hacktoberfest is all about &lt;strong&gt;celebrating open source&lt;/strong&gt;, and this hackathon perfectly captured that spirit. The CBIT edition — now in its 8th year — encouraged participants to collaborate on innovative solutions using modern technologies while contributing to the open-source ecosystem.&lt;/p&gt;

&lt;p&gt;Teams of &lt;strong&gt;3–5 members&lt;/strong&gt; worked virtually through &lt;strong&gt;Discord&lt;/strong&gt;, tackling real-world challenges in just 24 hours. Despite the distance and time zones, the sense of connection and creativity was palpable.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Judging Framework
&lt;/h2&gt;

&lt;p&gt;Every project was evaluated based on a clear, balanced set of criteria (total: 50 points):&lt;/p&gt;

&lt;div class="table-wrapper-paragraph"&gt;&lt;table&gt;
&lt;thead&gt;
&lt;tr&gt;
&lt;th&gt;Criterion&lt;/th&gt;
&lt;th&gt;Points&lt;/th&gt;
&lt;th&gt;What It Measured&lt;/th&gt;
&lt;/tr&gt;
&lt;/thead&gt;
&lt;tbody&gt;
&lt;tr&gt;
&lt;td&gt;Innovation &amp;amp; Creativity&lt;/td&gt;
&lt;td&gt;10&lt;/td&gt;
&lt;td&gt;Original ideas and new approaches&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Collaboration&lt;/td&gt;
&lt;td&gt;5&lt;/td&gt;
&lt;td&gt;Teamwork, communication, and shared contribution&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Implementation&lt;/td&gt;
&lt;td&gt;20&lt;/td&gt;
&lt;td&gt;Technical soundness, scalability, and efficiency&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Design&lt;/td&gt;
&lt;td&gt;5&lt;/td&gt;
&lt;td&gt;Usability and user experience&lt;/td&gt;
&lt;/tr&gt;
&lt;tr&gt;
&lt;td&gt;Presentation&lt;/td&gt;
&lt;td&gt;10&lt;/td&gt;
&lt;td&gt;Clarity, storytelling, and delivery&lt;/td&gt;
&lt;/tr&gt;
&lt;/tbody&gt;
&lt;/table&gt;&lt;/div&gt;

&lt;p&gt;As a judge, I was looking for that spark — projects that combined &lt;strong&gt;solid implementation with a clear purpose&lt;/strong&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  What I Saw and Learned
&lt;/h2&gt;

&lt;p&gt;Some projects were technically ambitious, pushing the boundaries of what’s possible in 24 hours. Others were beautifully simple, focusing on accessibility or social impact. A few used &lt;strong&gt;AI and automation&lt;/strong&gt; in creative ways that genuinely surprised me.&lt;/p&gt;

&lt;p&gt;But beyond the technology, what stood out was the &lt;strong&gt;teamwork&lt;/strong&gt;. Many participants were strangers before the event — yet they managed to code, design, and present together like long-time collaborators. That’s the magic of hackathons.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Judging Experience
&lt;/h2&gt;

&lt;p&gt;Our panel included engineers and leaders from &lt;strong&gt;Electronic Arts, JP Morgan Chase, Oracle, Deliveroo, ServiceNow, EY&lt;/strong&gt;, and more. Being part of such a diverse group gave every discussion depth. Each judge viewed “innovation” slightly differently — and that variety made our evaluations richer.&lt;/p&gt;

&lt;p&gt;I appreciated the effort teams put into their presentations. Even short demos told full stories — from the problem statement to the final prototype. That’s where creativity met clarity.&lt;/p&gt;

&lt;h2&gt;
  
  
  Key Takeaways
&lt;/h2&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Innovation isn’t always about technology.&lt;/strong&gt; Sometimes it’s about empathy — understanding who you’re helping and why.
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Good presentation matters.&lt;/strong&gt; The best projects told their stories with confidence and focus.
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Community is everything.&lt;/strong&gt; Open-source hackathons like this show how technology connects us beyond borders.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;As a developer, I’ve always believed in continuous learning. Judging this event reaffirmed that growth happens in many forms — whether you’re writing code or evaluating it.&lt;/p&gt;

&lt;h2&gt;
  
  
  Why Hackathons Matter More Than Ever
&lt;/h2&gt;

&lt;p&gt;In a world where remote collaboration is the new normal, hackathons remain one of the most human ways to innovate. They combine &lt;strong&gt;competition, creativity, and community&lt;/strong&gt; into one shared experience. For students and professionals alike, they’re the perfect place to experiment and grow.&lt;/p&gt;

&lt;p&gt;I’m grateful to have played a small role in this journey — to witness ideas take shape, to learn from participants, and to see firsthand how open source continues to inspire a new generation of builders.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Author:&lt;/strong&gt; &lt;br&gt;
&lt;em&gt;Servin Osmanov&lt;/em&gt;&lt;br&gt;&lt;br&gt;
Lead Software Engineer @ Anvaya Solutions Inc.&lt;br&gt;&lt;br&gt;
Judge at &lt;a href="https://cbit-hacktoberfest25.devpost.com/?ref_feature=challenge&amp;amp;ref_medium=discover#prizes" rel="noopener noreferrer"&gt;CBIT Hacktoberfest Hackathon 2025&lt;/a&gt;&lt;/p&gt;

</description>
      <category>hackathon</category>
      <category>webdev</category>
      <category>opensource</category>
      <category>ai</category>
    </item>
    <item>
      <title>Why Language Tech Matters: Developing AI Tools for Small Languages</title>
      <dc:creator>Servin Osmanov</dc:creator>
      <pubDate>Sat, 25 Oct 2025 15:45:27 +0000</pubDate>
      <link>https://forem.com/servin_osmanov/why-language-tech-matters-developing-ai-tools-for-small-languages-583h</link>
      <guid>https://forem.com/servin_osmanov/why-language-tech-matters-developing-ai-tools-for-small-languages-583h</guid>
      <description>&lt;p&gt;In a world where artificial intelligence is transforming how we communicate, the survival of small languages depends not just on cultural passion — but on technology.&lt;br&gt;&lt;br&gt;
While English, Chinese, or Spanish dominate the digital space, thousands of smaller languages remain digitally invisible. Without online presence, data, or digital tools, these languages risk extinction in the 21st century.  &lt;/p&gt;

&lt;p&gt;As a software engineer and cultural advocate, I’ve spent years developing digital tools for the &lt;strong&gt;Crimean Tatar&lt;/strong&gt; language — a Turkic minority language spoken by less than 300,000 people worldwide. Through this work, I’ve learned that the intersection of &lt;strong&gt;AI and linguistics&lt;/strong&gt; isn’t just a research topic; it’s a lifeline for cultural identity.&lt;/p&gt;




&lt;h2&gt;
  
  
  &lt;strong&gt;The Challenge: The Technology Gap for Small Languages&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;Mainstream AI models—from GPT-based chatbots to translation engines—are trained primarily on high-resource languages. This creates a serious imbalance: while global communication becomes easier for some, others are left behind.&lt;/p&gt;

&lt;p&gt;For small linguistic communities, the lack of:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;digital corpora,&lt;/li&gt;
&lt;li&gt;standardized spelling systems,&lt;/li&gt;
&lt;li&gt;and high-quality training datasets
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;makes it nearly impossible to integrate their languages into modern tools like voice assistants, translation services, or educational apps.&lt;/p&gt;

&lt;p&gt;When a language isn’t “machine-readable,” it risks becoming irrelevant in the digital world — even if it’s still spoken at home.&lt;/p&gt;




&lt;h2&gt;
  
  
  &lt;strong&gt;Our Journey: Building Tools for the Crimean Tatar Language&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;In 2018, our team launched a series of educational and linguistic tools under the &lt;strong&gt;Qırımtatar Lugatı&lt;/strong&gt; (Crimean Tatar Dictionary) project:  &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;a href="https://play.google.com/store/apps/details?id=com.anaurt.lugat" rel="noopener noreferrer"&gt;Android app on Google Play&lt;/a&gt;
&lt;/li&gt;
&lt;li&gt;&lt;a href="https://apps.apple.com/us/app/q%C4%B1r%C4%B1mtatar-lu%C4%9Fat%C4%B1/id1457493656" rel="noopener noreferrer"&gt;iOS app on the App Store&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Our mission was to give people an accessible and modern way to learn, use, and preserve the Crimean Tatar language.  &lt;/p&gt;

&lt;p&gt;The apps quickly grew into more than just dictionaries — they became &lt;strong&gt;living platforms&lt;/strong&gt; connecting speakers, learners, and educators. Yet, users wanted more: context-based search, on-the-fly translation, and natural voice pronunciation.&lt;br&gt;&lt;br&gt;
This inspired us to integrate &lt;strong&gt;artificial intelligence&lt;/strong&gt; to make our tools more flexible and intelligent.&lt;/p&gt;




&lt;h2&gt;
  
  
  &lt;strong&gt;The Next Step: Bringing AI to Minority Languages&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;We are now integrating &lt;strong&gt;AI and machine learning&lt;/strong&gt; into the Crimean Tatar dictionary project to expand its capabilities:  &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;🧠 &lt;strong&gt;AI-assisted translation:&lt;/strong&gt; dynamic, real-time translation between Crimean Tatar, Turkish, English, and Ukrainian.
&lt;/li&gt;
&lt;li&gt;🔊 &lt;strong&gt;Speech recognition and synthesis:&lt;/strong&gt; allowing users to hear natural pronunciation and practice correct intonation.
&lt;/li&gt;
&lt;li&gt;📖 &lt;strong&gt;Adaptive learning:&lt;/strong&gt; using AI to personalize vocabulary lessons based on user behavior and progress.
&lt;/li&gt;
&lt;li&gt;🪶 &lt;strong&gt;Data-driven NLP foundation:&lt;/strong&gt; building scalable open datasets that will support future chatbots, voice assistants, and translation systems.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;To achieve this, we’ve adopted and fine-tuned &lt;strong&gt;open-source speech synthesis models&lt;/strong&gt; such as:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;&lt;a href="https://huggingface.co/facebook/mms-tts-crh" rel="noopener noreferrer"&gt;facebook/mms-tts-crh&lt;/a&gt;&lt;/strong&gt; — part of Meta’s &lt;em&gt;Massively Multilingual Speech&lt;/em&gt; project, providing a solid baseline for Crimean Tatar TTS.
&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;&lt;a href="https://huggingface.co/spaces/robinhad/qirimtatar-tts" rel="noopener noreferrer"&gt;robinhad/qirimtatar-tts&lt;/a&gt;&lt;/strong&gt; — a community-driven model that helps us generate high-quality speech for educational and cultural content.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;By leveraging these models, we’re building an &lt;strong&gt;AI-powered TTS system&lt;/strong&gt; that brings Crimean Tatar audio resources to learners, teachers, and content creators for the first time.&lt;/p&gt;




&lt;h2&gt;
  
  
  &lt;strong&gt;Why This Matters&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;Language is more than communication — it’s collective memory.&lt;br&gt;&lt;br&gt;
When a language disappears, so does a unique worldview and cultural identity.  &lt;/p&gt;

&lt;p&gt;AI gives us a chance to reverse that process. With the right tools, small languages can become visible online, connect their communities, and survive in the digital era.&lt;/p&gt;

&lt;p&gt;Projects like &lt;strong&gt;Qırımtatar Lugatı&lt;/strong&gt; show that you don’t need a giant corporation to make an impact.&lt;br&gt;&lt;br&gt;
A small, passionate team with the right technical vision can give a digital voice — literally — to those who’ve been silent for too long.&lt;/p&gt;




&lt;h2&gt;
  
  
  &lt;strong&gt;How Developers and Researchers Can Help&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;If you’re a developer, linguist, or AI researcher, here are a few ways to make a difference:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Contribute to &lt;strong&gt;open-source datasets&lt;/strong&gt; for small or low-resource languages.
&lt;/li&gt;
&lt;li&gt;Support &lt;strong&gt;multilingual NLP and TTS&lt;/strong&gt; initiatives on Hugging Face or GitHub.
&lt;/li&gt;
&lt;li&gt;Collaborate with communities and educators who are digitizing endangered languages.
&lt;/li&gt;
&lt;li&gt;Help localize educational and cultural apps into underrepresented languages.
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Every dataset, every model, and every contribution brings us closer to a more linguistically inclusive AI ecosystem.&lt;/p&gt;




&lt;h2&gt;
  
  
  &lt;strong&gt;Looking Ahead&lt;/strong&gt;
&lt;/h2&gt;

&lt;p&gt;Our roadmap for 2025 includes:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Full integration of &lt;strong&gt;AI-powered translation&lt;/strong&gt; within the Qırımtatar Lugatı apps.
&lt;/li&gt;
&lt;li&gt;Implementation of &lt;strong&gt;voice features&lt;/strong&gt; using our fine-tuned TTS models.
&lt;/li&gt;
&lt;li&gt;Launching an &lt;strong&gt;open API&lt;/strong&gt; for developers who want to build tools for Crimean Tatar or similar minority languages.
&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Technology alone won’t preserve culture — but it can &lt;strong&gt;amplify&lt;/strong&gt; it.&lt;br&gt;&lt;br&gt;
For small languages like Crimean Tatar, &lt;strong&gt;AI isn’t just a tool — it’s hope.&lt;/strong&gt;&lt;/p&gt;




&lt;p&gt;&lt;strong&gt;About the Author&lt;/strong&gt;&lt;br&gt;&lt;br&gt;
&lt;em&gt;Servin Osmanov&lt;/em&gt; is a Senior Full-Stack Engineer and founder of the &lt;a href="https://qirim.online/" rel="noopener noreferrer"&gt;&lt;strong&gt;Qırım.Online&lt;/strong&gt;&lt;/a&gt; and &lt;a href="https://ana-yurt.com/" rel="noopener noreferrer"&gt;&lt;strong&gt;Ana-Yurt.Com&lt;/strong&gt;&lt;/a&gt; project — an initiative focused on preserving Crimean Tatar culture and language through modern technology.&lt;br&gt;&lt;br&gt;
LinkedIn: &lt;a href="https://www.linkedin.com/in/servin-osmanov/" rel="noopener noreferrer"&gt;linkedin.com/in/servin-osmanov&lt;/a&gt;&lt;br&gt;&lt;br&gt;
GitHub: &lt;a href="https://github.com/MrSerWin" rel="noopener noreferrer"&gt;github.com/MrSerWin&lt;/a&gt;&lt;/p&gt;

</description>
      <category>ai</category>
      <category>learning</category>
      <category>programming</category>
      <category>whisper</category>
    </item>
  </channel>
</rss>
