{
"lightbox_close": "Close",
"lightbox_next": "Next",
"lightbox_previous": "Previous",
"lightbox_error": "The requested content cannot be loaded. Please try again later.",
"lightbox_start_slideshow": "Start slideshow",
"lightbox_stop_slideshow": "Stop slideshow",
"lightbox_full_screen": "Full screen",
"lightbox_thumbnails": "Thumbnails",
"lightbox_download": "Download",
"lightbox_share": "Share",
"lightbox_zoom": "Zoom",
"lightbox_new_window": "New window",
"lightbox_toggle_sidebar": "Toggle sidebar"
}
Our lives are now packed with advanced devices, such as mobile gadgets, wearables, smart home appliances, telematics devices, and more.
Of all the features that make them advanced, the major one is the ability to understand user speech. Speaking into a device and telling it to do something are naturally easier and more satisfying than using input devices (like a keyboard and mouse) for the same purpose.
To help devices understand human speech, HMS Core ML Kit introduced the automatic speech recognition (ASR) service, to create a smoother human-machine interaction experience.
Service IntroductionASR can recognize and simultaneously convert speech (no longer than 60s) into text, by using industry-leading deep learning technologies. Boasting regularly updated algorithms and data, currently the service delivers a recognition accuracy of 95%+. The supported languages now are: Mandarin Chinese (including Chinese-English bilingual speech), English, French, German, Spanish, Italian, Arabic, Russian, Thai, Malay, Filipino, and Turkish.
Demo
Use CasesASR covers many fields spanning life and work, and enhances recognition capabilities of searching for products, movies, TV series, and music, as well as the capabilities for navigation services. When a user searches for a product in a shopping app through speech, this service recognizes the product name or feature in speech as text for search.
Similarly, when a user uses a music app, this service recognizes the song name or singer input by voice as text to search for the song.
On top of these, ASR can even contribute to driving safety: During driving — when users are not supposed to use their phone to, for example, search for a place — ASR allows them to speak out where they want to go and converts the speech into text for the navigation app which can then offer the search results to users.
Features
Real-time result output
Available options: with and without speech pickup UI
Endpoint detection: Start and end points of speech can be accurately located.
Silence detection: No voice packet is sent for silent parts.
Intelligent conversion of number formats: For example, when the speech is "year two thousand twenty-two", the text output by ASR will be "2022".
How to Integrate ML Kit?
For guidance about ML Kit integration, please refer to its official document. Also welcome to the HUAWEI Developers website, where you can find other resources for reference.
Related
Text Recognition with ML Kit
ML Kit gives developers the ability to implement text recognition into their apps. When using an API to develop your HMS-powered app, you'll have two different options. The text recognition API can be on-device or in-cloud. The on-device service will allow you to recognize Simplified Chinese, Japanese, Korean, and Latin-based languages (including English, Spanish, Portuguese, Italian, German, French, Russian, and special characters. The in-cloud API is more robust and allows you to recognize a wider variety of languages including Simplified Chinese, English, Spanish, Portuguese, Italian, German, French, Russian, Japanese, Korean, Polish, Finnish, Norwegian, Swedish, Danish, Turkish, Thai, Arabic, Hindi, and Indonesian.
The text recognition service is able to recognize text in both static images and dynamic camera streams with a host of APIs, which you can call synchronously or asynchronously to build your text recognition-enabled apps.
{
"lightbox_close": "Close",
"lightbox_next": "Next",
"lightbox_previous": "Previous",
"lightbox_error": "The requested content cannot be loaded. Please try again later.",
"lightbox_start_slideshow": "Start slideshow",
"lightbox_stop_slideshow": "Stop slideshow",
"lightbox_full_screen": "Full screen",
"lightbox_thumbnails": "Thumbnails",
"lightbox_download": "Download",
"lightbox_share": "Share",
"lightbox_zoom": "Zoom",
"lightbox_new_window": "New window",
"lightbox_toggle_sidebar": "Toggle sidebar"
}
Using the ML Kit demo APK, you can see this technology in action. The app is quick to accurately recognize any text your camera is pointed at. It takes less than a second for large text blocks to be converted into an actual text input for your phone. Translation features are also impressively fast, being able to read your words back to you in another language of your choice. This APK shows the extent to which this kit can be used, and makes development so much easier for these features.
How Developers are Implementing Text Recognition
There are many different ways that developers are taking advantage of ML Kit's text recognition. The ability to point your phone at some text and save it to your device opens many possibilities for great app ideas. You can use text recognition to quickly save the information off of a business card, translate text, create documents, and much more. Any situation where you can avoid requiring users to manually input text should be taken advantage of. This makes your app easier and quicker to use.
Whether a developer uses the on-device API or the in-cloud API depends on the needs of their app. The on-device API lets you add real-time processing of images from the camera stream. This means a user will be able to point their camera at some text, and the phone will be able to use ML Kit to recognize that text in real-time. The on-cloud API is better for high-accuracy text recognition from images and documents, but won't be able to complete real-time recognition from a camera.
Developer Resources
Huawei provides plenty of documentation and guides to help you get started with ML Kit's text recognition. You can get started with this guide here.
For all of the functions of ML Kit, refer to their service portal here.
For an overview of their APIs, browse the comprehensive resource library here.
You can also look at different ways that ML Kit can be implemented, by seeing a collection of sample codes here.
HMS ML Kit- Text to Speech
With HMS ML Kit Huawei ensures that developers have a simple way to implement text-to-speech features into their app. Text to speech has the ability to turn text content into a natural spoken voice. This service uses the deep neural network (DNN) synthesis mode and can be quickly integrated through the on-device SDK to generate audio data in real-time. It supports the download of offline models. In the current version, two standard male voices and six standard female voices are available. This service is available globally, to developers across the world.
Using the ML Kit demo APK, you can see this technology in action.
How Developers are Implementing Text to Speech
Some of the more common areas in which TTS is used can be found in broadcasting, news, voice navigation, and audio reading. Developers will be able to use this feature to allow their users to convert large amounts of text, into speech output. HMS TTS also works seamlessly with navigation data, which enables developers to create powerful navigation apps. ML Kit also understands how to synthesizes the voice segment into the navigation voice, so that navigation is more personalized.
{
"lightbox_close": "Close",
"lightbox_next": "Next",
"lightbox_previous": "Previous",
"lightbox_error": "The requested content cannot be loaded. Please try again later.",
"lightbox_start_slideshow": "Start slideshow",
"lightbox_stop_slideshow": "Stop slideshow",
"lightbox_full_screen": "Full screen",
"lightbox_thumbnails": "Thumbnails",
"lightbox_download": "Download",
"lightbox_share": "Share",
"lightbox_zoom": "Zoom",
"lightbox_new_window": "New window",
"lightbox_toggle_sidebar": "Toggle sidebar"
}
TTS is currently only available for users who have a Huawei device. Some of the limitations of this service include a cap on the amount of text that can be read, at 500 characters. Currently, TTS in French, Spanish, German, and Italian is deployed only in Asia, Africa, Latin America, and Europe. TTS depends on on-cloud APIs. During commissioning and usage, ensure that the device can access the Internet. Default specifications of the real-time output audio data are as follows: PCM mono, 16-bit depth, and 16 kHz audio sampling rate.
Developer Resources
Huawei provides plenty of documentation and guides to help you get started with ML Kit's Real-Time Transcription. You can get started with HMS ML Kit- Text to Speech with this guide here.
For all of the functions of ML Kit, refer to their service portal here.
For an overview of their APIs, browse the comprehensive resource library here.
You can also look at different ways that ML Kit can be implemented, by seeing a collection of sample codes here.
Grilling, stirring, baking, poaching… transforming ingredients into elegant dishes, cooking might be the most universal and enduring magic of human culture. But even so, making a simple pancake can be quite a challenge for a novice. With today's endless array of apps, there's no shortage of recipes at our disposal, for all tastes and diets. But if you're drawn in by the recipes of old and comfort foods, there's no app quite like Chefs'.
One-Touch Cookbook Digitalization
The creators behind Chefs' were a group of students from Warsaw University of Technology who are highly passionate about food, and were inspired in particularly by the dusty cookbooks in the basement. They hoped to create a handy culinary assistant that differs from any other cooking-related app on the market. Chefs', the fruit of their efforts, functions as a high-level digital cookbook for serious chefs and casual foodies alike.
In addition to offering a range of common services, such as searching and sharing functions, Chefs' also offers a unique one-touch image-based recipe generation feature, which enables you to take a snapshot or input an image from a cookbook, and be rewarded instantly with a clear, intuitive recipe on screen. This could spare you the endlessly tedious task of typing out recipes, and enables you to create a digital cookbook that can be shared and referenced with ease. Furthermore, the process of adding a recipe from an image is responsive and reliable, regardless of the network conditions.
{
"lightbox_close": "Close",
"lightbox_next": "Next",
"lightbox_previous": "Previous",
"lightbox_error": "The requested content cannot be loaded. Please try again later.",
"lightbox_start_slideshow": "Start slideshow",
"lightbox_stop_slideshow": "Stop slideshow",
"lightbox_full_screen": "Full screen",
"lightbox_thumbnails": "Thumbnails",
"lightbox_download": "Download",
"lightbox_share": "Share",
"lightbox_zoom": "Zoom",
"lightbox_new_window": "New window",
"lightbox_toggle_sidebar": "Toggle sidebar"
}
HUAWEI HiAI Open Platform: Endowing Your Apps with Intelligent Attributes
Chefs' leverages HUAWEI HiAI's state-of-the-art AI techniques, Text Image Super-Resolution and General Text Recognition, to accurately identify text blocks and fill in the blank spaces between with the corresponding content. Better yet, HUAWEI HiAI Engine is based entirely on the device, which means that the AI capabilities incur no server-related costs. The HUAWEI HiAI platform jumpstarted the development process, enabling Chefs' creators to forgo endless rounds of AI model training and the costly burden of server rents, enabling them to realize their vision – which resulted in the app winning the first prize at the Huawei AI CHALLENGE – Warsaw 2020.
If you can't wait to uncover the wisdom lurking in the yellowed, dog-eared pages of a family cookbook, or yearning to collect all of your recipes in a single place, you'd be remiss not to check out Chefs' on HUAWEI AppGallery.
Fore more information,please visit DevHub
cr. Amber - Brilliant Recipes Preserved for Posterity, with Chefs' Powered by HUAWEI HiAI
Notebloc is a Spanish app that scans, saves, and shares notes, documents, receipts, drawings, photos, and images of any type. It allows users to crop documents or images, and automatically corrects misaligned pages. Notebloc, which is now available worldwide and supports over 30 languages, has already been downloaded by more than 7 million users around the world.
{
"lightbox_close": "Close",
"lightbox_next": "Next",
"lightbox_previous": "Previous",
"lightbox_error": "The requested content cannot be loaded. Please try again later.",
"lightbox_start_slideshow": "Start slideshow",
"lightbox_stop_slideshow": "Stop slideshow",
"lightbox_full_screen": "Full screen",
"lightbox_thumbnails": "Thumbnails",
"lightbox_download": "Download",
"lightbox_share": "Share",
"lightbox_zoom": "Zoom",
"lightbox_new_window": "New window",
"lightbox_toggle_sidebar": "Toggle sidebar"
}
Notebloc's core functions are centered around documents. The integration of the HMS Core Pencil Engine into the Notebloc app offers specialized features such as brush effects, handwriting editing, stroke estimate, smart shape, and double-tapping. These advanced tools provide a superior handwriting experience. Now, users can effortlessly edit documents, by using the marker to annotate, mark up, and add notes to a file, and they can also unleash their creativity by adding sketches or diagrams. This is how Huawei's Pencil Engine allows Notebloc to bring users' best ideas to life.
Notebloc also integrates the HMS Core ML Kit text recognition service, which enables the app to accurately identify and extract text from images of receipts, business cards, and documents, and provide precise, and structured transcription of important information in text, greatly improving user satisfaction.
Teamwork Timeline:2013:Notebloc was founded in Barcelona, Spain
2021:In September, team meetings were held regarding co-development between Notebloc and Huawei. In November, the project began. Huawei's HMS Core DTSE team helped Notebloc's developers overcome difficulties, such as a lack of test devices and insufficient sample documents.
2022:In January, HMS Core was successfully integrated into the Notebloc app, and became available on the HUAWEI AppGallery.
Customer feedback:
STEM Alliance reported that Notebloc, one of the first apps in Europe to integrate Pencil Engine, was able to provide users with substantially improved note-taking services by working with HMS Core. More users can now access and use a myriad of editing tools, and can easily, and securely, scan and share documents.
The Notebloc team confirmed its intention to integrate other HMS Core capabilities in order to attract app users and increase monetization in the future.
To learn more, please visit:
>> Reddit to join developer discussions
>> GitHub to download the sample code
>> Stack Overflow to solve integration problems
Follow our official account for the latest HMS Core-related news and updates.
Since 1839 when Louis Daguerre invented the daguerreotype (the first publicly available photographic process), new inventions have continued to advance photography. Its spike reached a record high where people were able to record experiences through photos, anytime and anywhere. However, it is a shame that many early photos existed in only black and white.
HMS Core Video Editor Kit provides the AI color function that can liven up such photos, intelligently adding color to black-and-white images or videos to endow them with a more contemporary feel.
{
"lightbox_close": "Close",
"lightbox_next": "Next",
"lightbox_previous": "Previous",
"lightbox_error": "The requested content cannot be loaded. Please try again later.",
"lightbox_start_slideshow": "Start slideshow",
"lightbox_stop_slideshow": "Stop slideshow",
"lightbox_full_screen": "Full screen",
"lightbox_thumbnails": "Thumbnails",
"lightbox_download": "Download",
"lightbox_share": "Share",
"lightbox_zoom": "Zoom",
"lightbox_new_window": "New window",
"lightbox_toggle_sidebar": "Toggle sidebar"
}
In addition to AI color, the kit also provides other AI-empowered capabilities, such as allowing your users to copy a desired filter, track motions, change hair color, animate a picture, and mask faces.
In terms of input and output support, Video Editor Kit allows multiple images and videos to be imported, which can be flexibly arranged and trimmed, and allows videos of up to 4K and with a frame rate up to 60 fps to be exported.
Useful in Various Scenarios
Video Editor Kit is ideal for numerous application scenarios, to name a few:
Video editing: The kit helps accelerate video creation by providing functions such as video clipping/stitching and allowing special effects/music to be added.
Travel: The kit enables users to make vlogs on the go to share their memories with others.
Social media: Functions like video clipping/stitching, special effects, and filters are especially useful for social media app users, and are a great way for them to spice up videos.
E-commerce: Product videos with subtitles, special effects, and background music allow products to be displayed in a more intuitive and immersive way.
Flexible Integration Methods
Video Editor Kit can now be integrated via its:
UI SDK, which comes with a product-level UI for straightforward integration.
Fundamental capability SDK, which offers hundreds of APIs for fundamental capabilities, including the AI-empowered ones. The APIs can be integrated as needed.
Both of the SDKs serve as a one-stop toolkit for editing videos, providing functions including file import, editing, rendering, output, and material management. Integrating either of the SDKs allows you to access the kit's powerful capabilities.
These capabilities enable your users to restore early photos and record life experiences. Check out the official documentation for this great Video Editor Kit, to know more about how it can help you create a mobile life recorder.