<?xml version="1.0" encoding="UTF-8"?>
<!-- AUTOGENERATED FILE. DO NOT EDIT. -->
<feed xmlns="http://www.w3.org/2005/Atom">
  <id>tag:google.com,2016:speech-release-notes</id>
  <title>Speech-to-Text - Release notes</title>
  <link rel="self" href="https://docs.cloud.google.com/feeds/speech-release-notes.xml"/>
  <author>
    <name>Google Cloud Platform</name>
  </author>
  <updated>2025-11-13T00:00:00-08:00</updated>

  <entry>
    <title>November 13, 2025</title>
    <id>tag:google.com,2016:speech-release-notes#November_13_2025</id>
    <updated>2025-11-13T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#November_13_2025"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Speech-to-Text has just launched <code>chirp_3</code> in public preview for the regions <code>asia-south1</code>, <code>europe-west2</code>, <code>europe-west3</code> and <code>northamerica-northeast1</code>.</p>
<p>For more information about the Chirp 3 model, see <a href="https://docs.cloud.google.com/speech-to-text/docs/models/chirp-3">Chirp 3 Transcription: Enhanced multilingual accuracy</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>October 13, 2025</title>
    <id>tag:google.com,2016:speech-release-notes#October_13_2025</id>
    <updated>2025-10-13T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#October_13_2025"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Speech-to-Text is excited to announce the General Availability (GA) of the Chirp 3: Transcription, the latest generation of Google's multilingual Automatic Speech Recognition (ASR)-specific generative model, delivering state-of-the-art ASR accuracy and multilingual capabilities. Available exclusively in the Speech-to-Text API V2, Chirp 3 delivers significant enhancements in transcription accuracy and speed over previous versions. 
Under the new <code>chirp_3</code> model identifier, you can now leverage powerful new capabilities, including speaker diarization to identify different speakers and automatic language detection for multilingual audio. The model supports all major recognition methods —<code>StreamingRecognize</code>, <code>Recognize</code>, and <code>BatchRecognize</code>- making it suitable for both real-time and batch processing. Chirp 3 also offers advanced features such as speech adaptation for custom vocabularies and a built-in denoiser to improve results from noisy audio. </p>
<p>To explore the new Chirp 3: Transcription model's capabilities and learn how to leverage its full potential, please visit our <a href="https://docs.cloud.google.com/speech-to-text/docs/models/chirp-3">updated documentation page</a>. </p>
]]>
    </content>
  </entry>

  <entry>
    <title>August 29, 2025</title>
    <id>tag:google.com,2016:speech-release-notes#August_29_2025</id>
    <updated>2025-08-29T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#August_29_2025"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Speech-to-Text has just launched <code>chirp_3</code> in Public Preview. With this Public Preview, Chirp 3: Transcription we are now expanding language transcription in more than 85+ languages and locales, in addition to StreamingRecognize and SyncRecognize requests for real-time and short-form audio. Under the <code>chirp_3</code> model flag, you can experience significant improvements in accuracy and speed, and leverage powerful features like Speaker Diarization and language-agnostic transcription. </p>
<p>To explore the new Chirp 3: Transcription model's capabilities and learn how to leverage its full potential, please visit our <a href="https://cloud.google.com/speech-to-text/v2/docs/chirp_3-model">official documentation page</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>April 11, 2025</title>
    <id>tag:google.com,2016:speech-release-notes#April_11_2025</id>
    <updated>2025-04-11T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#April_11_2025"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Speech-to-Text has launched <code>chirp_3</code> in Private Preview. Chirp 3: Transcription is the latest generation of Google's multilingual Automatic Speech Recognition (ASR)-specific generative models that further enhances its ASR accuracy and multilingual capabilities. Under the new <code>chirp_3</code> model flag, you can experience significant improvements in accuracy and speed, and leverage powerful new features like Speaker Diarization and language-agnostic transcription. Chirp 3 supports <code>BatchRecognize</code> requests within the Speech-to-Text v2 API, making it ideal for transcribing long-form audio.</p>
<p>To explore the new Chirp 3: Transcription model's capabilities and to learn how to leverage its full potential, please visit our <a href="https://cloud.google.com/speech-to-text/v2/docs/chirp_3-model">official documentation page</a>. To gain access to this Private Preview, please <a href="https://cloud.google.com/contact">contact our sales team</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>January 27, 2025</title>
    <id>tag:google.com,2016:speech-release-notes#January_27_2025</id>
    <updated>2025-01-27T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#January_27_2025"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Speech-to-Text is generally available (GA) in the Chirp 2 model in <code>asia-southeast1</code>, <code>us-central1</code>, and <code>europe-west4</code>.</p>
<p>For more information about the Chirp 2 model, see <a href="https://docs.cloud.google.com/speech-to-text/v2/docs/chirp_2-model?_gl=1*la7eog*_ga*NzAyNzc4MTMxLjE3MzY5NjgxMTA.*_ga_4LYFWVHBEB*MTczODAxMTAzMy45My4xLjE3MzgwMTE2OTMuMC4wLjA.">Chirp 2: Enhanced multilingual accuracy</a>. For code samples, see <a href="https://github.com/GoogleCloudPlatform/generative-ai/blob/main/audio/speech/getting-started/get_started_with_chirp_2_sdk.ipynb">Get started with Chirp 2 using Speech-to-Text V2 SDK in GitHub</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>October 07, 2024</title>
    <id>tag:google.com,2016:speech-release-notes#October_07_2024</id>
    <updated>2024-10-07T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#October_07_2024"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Speech-to-Text has updated the Generally Available Chirp 2 model, further enhancing its ASR accuracy and multilingual capabilities. Under the existing <code>chirp_2</code> model flag, you can experience significant improvements in accuracy and speed, as well as support for word-level timestamps, model adaptation, and speech translation. Finally, Chirp 2 can support Streaming Recognizer requests, in addition to the already supported Sync and Batch Recognition requests, allowing its use in realtime applications.</p>
<p>Explore the new chirp_2 model's capabilities and learn how to leverage its full potential by visiting our <a href="https://docs.cloud.google.com/speech-to-text/v2/docs/chirp_2-model">updated documentation</a> and <a href="https://github.com/GoogleCloudPlatform/generative-ai/blob/main/audio/speech/getting-started/get_started_with_chirp_2_sdk.ipynb">tutorials</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>January 09, 2024</title>
    <id>tag:google.com,2016:speech-release-notes#January_09_2024</id>
    <updated>2024-01-09T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#January_09_2024"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p><a href="https://docs.cloud.google.com/speech-to-text/docs/adaptation-model">Model adaptation</a> is now available for <code>latest_long</code> models in 13 languages. Also, its quality was substantially improved for <code>latest_short</code> models. To determine whether this feature is available for your language, see <a href="https://docs.cloud.google.com/speech-to-text/docs/speech-to-text-supported-languages">Language support</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>January 08, 2024</title>
    <id>tag:google.com,2016:speech-release-notes#January_08_2024</id>
    <updated>2024-01-08T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#January_08_2024"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Speech-to-Text has launched a new model, named <code>chirp_telephony</code> to bring the accuracy gains of our <code>chirp</code> model to telephony-specific use cases. The new model is a fine-tuned version of our very successful <code>chirp</code> model, based on the Universal large Speech Model(USM) architecture, on audio that originated from a phone call typically recorded at an 8 kHz sampling rate. For more information, see <a href="https://docs.cloud.google.com/speech-to-text/v2/docs/speech-to-text-supported-languages">Speech-to-Text supported languages</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>November 06, 2023</title>
    <id>tag:google.com,2016:speech-release-notes#November_06_2023</id>
    <updated>2023-11-06T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#November_06_2023"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Speech-to-Text has launched two models, named <code>telephony</code> and <code>telephony_short</code>. The two models are customized to recognize audio that originates from a phone call and corresponds to the most recent versions of the existing <code>phone_call</code> model. For more information, see <a href="https://docs.cloud.google.com/speech-to-text/docs/speech-to-text-supported-languages">Speech-to-Text supported languages</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>February 07, 2023</title>
    <id>tag:google.com,2016:speech-release-notes#February_07_2023</id>
    <updated>2023-02-07T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#February_07_2023"/>
    <content type="html"><![CDATA[<h3>Announcement</h3>
<p>We are removing <code>SpeechContext.strength</code> field within the next 4
weeks, because it has been deprecated and unused for more than a year. The
documentation doesn't have references to this field anymore, and the clients aren't supposed to use it.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>November 11, 2022</title>
    <id>tag:google.com,2016:speech-release-notes#November_11_2022</id>
    <updated>2022-11-11T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#November_11_2022"/>
    <content type="html"><![CDATA[<h3>Change</h3>
<p>Speech-to-Text has updated its pricing policy. Enhanced models are no longer priced differently than standard models. Usage of all models will be reported to and priced like standard models. Also, all Cloud Speech-to-Text requests will now be rounded up to the nearest 1 second, with no minimum audio length (requests were previously rounded up to the nearest 15 seconds). See the <a href="https://cloud.google.com/speech-to-text/pricing">Pricing</a> page for details.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>October 03, 2022</title>
    <id>tag:google.com,2016:speech-release-notes#October_03_2022</id>
    <updated>2022-10-03T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#October_03_2022"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Speaker Diarization is now available for "Latest" models in en-US. This feature recognizes multiple speakers in the same audio clip. Latest models use a new model for diarization from previous models. For more information see <a href="https://cloud.google.com/speech-to-text/docs/multiple-voices#speaker_diarization">Speaker Diarization</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>April 21, 2022</title>
    <id>tag:google.com,2016:speech-release-notes#April_21_2022</id>
    <updated>2022-04-21T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#April_21_2022"/>
    <content type="html"><![CDATA[<h3>Change</h3>
<p>"Latest" models are available in more than 20 languages. These models employ new end-to-end machine learning techniques and can improve the accuracy of your recognized speech. For more information see <a href="https://docs.cloud.google.com/speech-to-text/docs/latest-models">Latest models</a>.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>November 08, 2021</title>
    <id>tag:google.com,2016:speech-release-notes#November_08_2021</id>
    <updated>2021-11-08T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#November_08_2021"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Speech-to-Text has launched two new medical speech models, which are tailored for recognition of words that are common in medical settings. See the <a href="https://docs.cloud.google.com/speech-to-text/docs/medical-models">medical models documentation</a> for more details.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>July 21, 2021</title>
    <id>tag:google.com,2016:speech-release-notes#July_21_2021</id>
    <updated>2021-07-21T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#July_21_2021"/>
    <content type="html"><![CDATA[<h3>Announcement</h3>
<p>Speech-to-Text has launched a GA version of the Spoken Emoji and Spoken Punctuation features. See the <a href="https://docs.cloud.google.com/speech-to-text/docs/spoken-punctuation-emojis">documentation</a> for details.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>June 28, 2021</title>
    <id>tag:google.com,2016:speech-release-notes#June_28_2021</id>
    <updated>2021-06-28T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#June_28_2021"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>The Speech-to-Text now supports multi-region endpoints as a GA feature. See the <a href="https://docs.cloud.google.com/speech-to-text/docs/endpoints">multi-region endpoints</a> documentation for more information.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>May 24, 2021</title>
    <id>tag:google.com,2016:speech-release-notes#May_24_2021</id>
    <updated>2021-05-24T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#May_24_2021"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Speech-to-Text now supports Spoken Punctuation and Spoken Emoji as Preview features. See the <a href="https://docs.cloud.google.com/speech-to-text/docs/spoken-punctuation-emojis">documentation</a> for details.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>May 07, 2021</title>
    <id>tag:google.com,2016:speech-release-notes#May_07_2021</id>
    <updated>2021-05-07T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#May_07_2021"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>The Speech-to-Text <a href="https://docs.cloud.google.com/speech-to-text/docs/adaptation">model adaptation</a> feature is now a GA feature. See the <a href="https://docs.cloud.google.com/speech-to-text/docs/adaptation-model">model adaptation concepts page</a> for more information about using this feature.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>March 23, 2021</title>
    <id>tag:google.com,2016:speech-release-notes#March_23_2021</id>
    <updated>2021-03-23T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#March_23_2021"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Speech-to-Text now allows you to upload your longrunning transcription results directly into a Cloud Storage bucket. See the <a href="https://docs.cloud.google.com/speech-to-text/docs/async-recognize#uploading_your_transcription_results_to_a_gcs_bucket">asynchronous speech recognition documentation</a> for more details.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>March 15, 2021</title>
    <id>tag:google.com,2016:speech-release-notes#March_15_2021</id>
    <updated>2021-03-15T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#March_15_2021"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Speech-to-Text has launched the <a href="https://docs.cloud.google.com/speech-to-text/docs/adaptation-model">Model Adaptation feature</a>. You can now create
<a href="https://docs.cloud.google.com/speech-to-text/docs/adaptation-model#custom_classes">custom classes</a> and build <a href="https://docs.cloud.google.com/speech-to-text/docs/adaptation">phrase sets</a> to improve your transcription results.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>January 26, 2021</title>
    <id>tag:google.com,2016:speech-release-notes#January_26_2021</id>
    <updated>2021-01-26T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#January_26_2021"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Speech-to-Text now supports regional EU and US endpoints. See the <a href="https://docs.cloud.google.com/speech-to-text/docs/endpoints">multi-region endpoints</a> documentation for more information.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>August 25, 2020</title>
    <id>tag:google.com,2016:speech-release-notes#August_25_2020</id>
    <updated>2020-08-25T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#August_25_2020"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Speech-to-Text has launched the new <a href="https://docs.cloud.google.com/speech-to-text/on-prem">On-Prem API</a>. Speech-to-Text On-Prem enables easy integration of Google speech recognition technologies into your on-premises solution.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>March 05, 2020</title>
    <id>tag:google.com,2016:speech-release-notes#March_05_2020</id>
    <updated>2020-03-05T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#March_05_2020"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Cloud Speech-to-Text now supports seven new <a href="https://docs.cloud.google.com/speech-to-text/docs/languages">languages</a>: Burmese, Estonian, Uzbek, Punjabi, Albanian, Macedonian, and Mongolian.</p>
<h3>Feature</h3>
<p>The <a href="https://docs.cloud.google.com/speech-to-text/docs/multiple-voices">speaker diarization</a>, <a href="https://docs.cloud.google.com/speech-to-text/docs/automatic-punctuation">automatic punctuation</a>, <a href="https://docs.cloud.google.com/speech-to-text/docs/boost">speech adaptation boost</a>, and <a href="https://docs.cloud.google.com/speech-to-text/docs/enhanced-models">enhanced telephony model</a> features are now available for new languages. See the supported <a href="https://docs.cloud.google.com/speech-to-text/docs/languages">languages page</a> for a complete list.</p>
<h3>Change</h3>
<p><a href="https://docs.cloud.google.com/speech-to-text/docs/class-tokens">Class tokens</a> are now available for general use. You can use class tokens with <a href="https://docs.cloud.google.com/speech-to-text/docs/speech-adaptation#classes">speech adaptation</a> to help the model recognize concepts in your recorded audio data.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>November 26, 2019</title>
    <id>tag:google.com,2016:speech-release-notes#November_26_2019</id>
    <updated>2019-11-26T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#November_26_2019"/>
    <content type="html"><![CDATA[<h3>Change</h3>
<p><a href="https://docs.cloud.google.com/speech-to-text/docs/automatic-punctuation">Automatic punctuation</a> is now available for general use. Cloud Speech-to-Text can insert punctuation into transcription results, including commas, periods, and question marks.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>July 23, 2019</title>
    <id>tag:google.com,2016:speech-release-notes#July_23_2019</id>
    <updated>2019-07-23T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#July_23_2019"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Cloud Speech-to-Text has several <a href="https://docs.cloud.google.com/speech-to-text/docs/endless-streaming-tutorial">endless streaming tutorials</a> that demonstrate how to transcribe an infinite audio stream.</p>
<h3>Feature</h3>
<p>You can now use <a href="https://docs.cloud.google.com/speech-to-text/docs/context-strength">speech adaptation</a> to provide 'hints' to Cloud Speech-to-Text when it performs speech recognition. This feature is now in beta.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>June 18, 2019</title>
    <id>tag:google.com,2016:speech-release-notes#June_18_2019</id>
    <updated>2019-06-18T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#June_18_2019"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Cloud Speech-to-Text now has expanded to a 5 minute limit for
<a href="https://docs.cloud.google.com/speech-to-text/docs/streaming-recognize">streaming recognition</a>.
To use streaming recognition with the 5 minute limit, you must use the
<a href="https://docs.cloud.google.com/speech-to-text/docs/reference/rpc/google.cloud.speech.v1p1beta1#speech"><code>v1p1beta1</code></a>
API version.</p>
<h3>Feature</h3>
<p>Cloud Speech-to-Text now supports transcription of <a href="https://docs.cloud.google.com/speech-to-text/docs/encoding">MP3 encoded audio data</a>.
As this feature is in beta, you must use the
<a href="https://docs.cloud.google.com/speech-to-text/docs/reference/rpc/google.cloud.speech.v1p1beta1#speech"><code>v1p1beta1</code></a>
API version.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>April 04, 2019</title>
    <id>tag:google.com,2016:speech-release-notes#April_04_2019</id>
    <updated>2019-04-04T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#April_04_2019"/>
    <content type="html"><![CDATA[<h3>Deprecated</h3>
<p>The <code>v1beta</code> version of the service is no longer available for use. You must migrate your solutions to either the
<a href="https://docs.cloud.google.com/speech-to-text/docs/reference/rpc/google.cloud.speech.v1"><code>v1</code></a> or
<a href="https://docs.cloud.google.com/speech-to-text/docs/reference/rpc/google.cloud.speech.v1p1beta1"><code>v1p1beta1</code></a> version of the API.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>February 20, 2019</title>
    <id>tag:google.com,2016:speech-release-notes#February_20_2019</id>
    <updated>2019-02-20T00:00:00-08:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#February_20_2019"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p><a href="https://docs.cloud.google.com/speech-to-text/docs/data-logging">Data logging</a>
is now available for general use. When you enable data logging,
you can reduce the cost of using Cloud Speech-to-Text by allowing Google to log
your data in order to improve the service.</p>
<h3>Change</h3>
<p>Using <a href="https://docs.cloud.google.com/speech-to-text/docs/enhanced-models">enhanced models</a>
no longer requires you to
opt-in for data logging. Enhanced models are available for use by any transcription requests
for a different
<a href="https://cloud.google.com/speech-to-text/pricing">price</a>
as standard models.</p>
<h3>Feature</h3>
<p><a href="https://docs.cloud.google.com/speech-to-text/docs/enhanced-models">Enhanced models</a>
are now available for general use.
Using enhanced models can improve audio transcription results.</p>
<h3>Feature</h3>
<p><a href="https://docs.cloud.google.com/speech-to-text/docs/transcription-model">Selecting a transcription model</a>
is now available for general use.
You can select different speech recognition models when you send a request to
Cloud Speech-to-Text, including a model optimized for transcribing audio data
from video files.</p>
<h3>Feature</h3>
<p>You can now include more details about your audio source files in transcription requests to Cloud Speech-to-Text in the form of <a href="https://docs.cloud.google.com/speech-to-text/docs/samples/speech-transcribe-recognition-metadata-beta">recognition metadata</a>, which can improve the results of the speech recognition. This feature  is now available for general use.</p>
<h3>Feature</h3>
<p>Cloud Speech-to-Text can transcribe audio data that includes
<a href="https://docs.cloud.google.com/speech-to-text/docs/multi-channel">multiple channels</a>.
This feature is now available for general use.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>July 24, 2018</title>
    <id>tag:google.com,2016:speech-release-notes#July_24_2018</id>
    <updated>2018-07-24T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#July_24_2018"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Cloud Speech-to-Text provides
<a href="https://docs.cloud.google.com/speech-to-text/docs/word-confidence">word-level confidence</a>
Developers can use this feature to get the degree of confidence on a word-by-word level. This feature is in Beta.</p>
<h3>Feature</h3>
<p>Cloud Speech-to-Text can automatically detect the language used in an audio file.
To use this feature, developers must specify
<a href="https://docs.cloud.google.com/speech-to-text/docs/multiple-languages">alternative languages</a>
in their transcription request. This feature is in Beta.</p>
<h3>Feature</h3>
<p>Cloud Speech-to-Text can
<a href="https://docs.cloud.google.com/speech-to-text/docs/multiple-voices">identify different speakers</a>
present in an audio file. This feature is in Beta.</p>
<h3>Feature</h3>
<p>Cloud Speech-to-Text can transcribe audio data that includes
<a href="https://docs.cloud.google.com/speech-to-text/docs/multi-channel">multiple channels</a>.
This feature is in Beta.</p>
]]>
    </content>
  </entry>

  <entry>
    <title>April 09, 2018</title>
    <id>tag:google.com,2016:speech-release-notes#April_09_2018</id>
    <updated>2018-04-09T00:00:00-07:00</updated>
    <link rel="alternate" href="https://docs.cloud.google.com/speech-to-text/docs/release-notes#April_09_2018"/>
    <content type="html"><![CDATA[<h3>Feature</h3>
<p>Cloud Speech-to-Text now provides
<a href="https://docs.cloud.google.com/speech-to-text/docs/data-logging">data logging</a> and
<a href="https://docs.cloud.google.com/speech-to-text/docs/enhanced-models">enhanced models</a>.
Developers that want to take advantage of the enhanced speech recognition models can
<a href="https://docs.cloud.google.com/speech-to-text/docs/data-logging-terms">opt-in for data logging</a>.
This feature is in Beta.</p>
<h3>Feature</h3>
<p>Cloud Speech-to-Text can insert
<a href="https://docs.cloud.google.com/speech-to-text/docs/automatic-punctuation">punctuation</a>
into transcription results, including commas, periods, and question marks. This feature is in Beta.</p>
<h3>Feature</h3>
<p>You can now select different speech recognition models when you send
a request to Cloud Speech-to-Text, including a model optimized for
<a href="https://docs.cloud.google.com/speech-to-text/docs/video-model">transcribing audio from video files</a>.
This feature is in Beta.</p>
<h3>Feature</h3>
<p>You can now include more details about your audio
source files in transcription requests to Cloud Speech-to-Text in the form of
<a href="https://docs.cloud.google.com/speech-to-text/docs/samples/speech-transcribe-recognition-metadata-beta">recognition metadata</a>,
which can improve the results of the speech recognition. This feature is in Beta.</p>
]]>
    </content>
  </entry>

</feed>
