Google’s on-device ML Kit recently introduced another API for barcode scanning. It can read and scan almost a dozen different types of barcodes including Codabar, Code 39, Code 93, EAN-8, EAN-13, QR code, PDF417, and more.
The barcode scanning API also tells you the scanned format and it will automatically detect the formats. The other main feature of this API is that you can use it in your application without a network connection and it will support the different orientation.
In this article, we’ll do just that on Android with the help of Google ML Kit’s Barcode Scanning API.
By now, you are probably familiar with the concept of “smart reply” features. In many messaging applications nowadays, you’ll see a list of suggested replies when you receive a message. And this makes it easier to respond without thinking about what you’re going to say—especially when the reply is obvious.
You can easily add Smart Reply to your mobile apps. like chat-app, Reviews, with the help of Google’s API, via its on-device machine learning framework ML Kit, which allows you to easily add this capability to your projects.
In this article, we’ll do just that on Android with the help…
Creating accurate machine learning models capable of identifying multiple faces (or other target objects) in a single image remains a core challenge in computer vision, but now with the advancement of deep learning and computer vision models for mobile, you can much more easily detect faces in an image.
In this article, we’ll do just that on Android with the help of Google ML Kit’s Face Detection API.
Face detection is a computer vision task with which you can detect faces in an image—it also detects the various part of the face, known as landmarks. …
Shake is a bug reporting SDK for iOS and Android apps. It’s a powerful tool that easily compares to other bug reporting tools like Instabug or Crashlytics. In this article, I’ll explore some of the main features of the Shake SDK, it’s benefits, and how you can integrate it into your Android Project.
With Shake SDK added to your app, your testers can report any bug in under 5 seconds. It will take all the extra info you as a developer need and send it to the Shake web dashboard:
Shake SDK automatically captures the following information:
Nowadays, language detection is very popular (especially with machine learning), and mobile apps that use it are widely popular in every part of the world, with different users speaking different languages. Language identification can easily help you understand your users’ languages and personalize your app based on them.
Language detection is essentially a technique/science that allows us to automatically identify the language of a given text, be it English, Chinese, or many others. We can use machine learning for this kind of identification —and we can even do this inside mobile apps!
In this article, we’ll do just that, with…
Firebase Crashlytics is a real-time crash reporting tool. It helps by automatically collecting, analyzing, and organizing your crash reports.
It also helps you understand which issues are most important, so you can priorities those first and keep your users happy. It combines a real-time dashboard and alerts to keep you aware of your newest issues and any sudden changes instability.
The first step we need to take is to create a Firebase project and add it to our Android app. At this point, I’ve already created a Firebase project, added it to a new project, and also included the Google…
In the world of computer science, there are many programming languages, and no single language is superior to another. In other words, each language is best suited to solve certain problems, and in fact, there is often no one best language to choose for a given programming project. For this reason, it is important for students who wish to develop software or to solve interesting problems through code to have strong computer science fundamentals that will apply across any programming language.
Programming languages tend to share certain characteristics in how they function, for…
Creating accurate machine learning models capable of identifying multiple objects in a single image remains a core challenge in computer vision, but now with the advancement of deep learning and computer vision models for mobile, you can easily detect target objects from an image on Android.
In this article, we’ll do just that with the help of Firebase ML Kit’s Object Detection API.
Object detection is a computer vision technique for identifying and locating instances of objects within images or videos. It provides much better understanding of an image as a whole when compared to visual recognition (i.e. …
Nowadays In Android applications, we often have seen custom fonts throughout the application, as the custom font in detailing, the custom font in the toolbar, custom font everywhere in the application. Which makes UI more appropriate and appealing.
In this article, we will learn how we can customize the toolbar font, how to change the Toolbar font family with the help of a theme.
Originally published at https://danishamjad.com on June 2, 2020.
Before directly jumping towards custom toolbar font, first, see how we can easily change the font family for
TextView in Android.
We just need to add the
Slow build speed is a huge momentum buster. It’s kind of like driving down the road and constantly hitting speed bump after speed bump. Build speed is critical to your productivity, and over the past few years, the Android team has consistently focused on improving build speed performance.
After implementing the tips in this article, we’ll be able to speed up development build times by 3x, 4x—sometimes even up to 10x. All by simply applying a set of best practices.
To help contextualize and show the impact of these tips, I’m going to use Google’s Santa Tracker Android project as…