Does azure conversation transcription service is available in "python"? - python

Problem statement: Need to transcript the speech to text in real-time and distinguish the user
as speaker 1 and speaker 2 using azure cognitive speech service.
Until now I explore the documentation of azure regarding conversation transcription which provides the sample code for Javascript and C#link for the documentationbut I was not able to find the sample code in python so does that means azure's this service is not available in python?

Does azure conversation transcription service is available in "python"?
No, at present Conversation Transcription SDK does not support Python language.
Conversation Transcription SDK supports only c# and javascript and is only available in few regions like centralus, eastasia, eastus, westeurope.
You can reach Microsoft here for support.

Related

How to publish message to Google Cloud MQTT topic?

Using Google Cloud Functions with Python, I want to publish a message to MQTT topic.
I use pubsub_v1 PublisherClient, and after invoking cloud function and looking into logs it looks like an "google.api_core.exceptions.InvalidArgument: 400 Invalid resource name given". I understand that happens because my topic_name contains "/", because I want to use more levels/subtopics, something like this : "devices/raspberrypi/events/msg/device/3/status". I have "msg" topic created, but can't create or publish to any subtopics.
I didn't find any information about subtopics in official documentation or anywhere else. Thanks in advance.
The Cloud Pub/Sub client library is only for interacting with Google Cloud Pub/Sub directly. If you are using MQTT topics, it sounds like you are trying to use the Cloud IoT Core MQTT bridge. In that case, you need to use an MQTT client that is configured to authenticate the device with Cloud IoT Core and use it to publish the telemetry events. See the Cloud IoT guide to publishing.
The subtopics for events should have their own topics associated with them, but they are not going to be accessible using the subtopic name via the Cloud Pub/Sub library. To use the Cloud Pub/Sub library, you will have to use the name of the Cloud Pub/Sub topic. The topic would be the one you associated with the MQTT subtopic in the Cloud Console in the "Additional topic" section:

Working with Watson Assistant / Watson Conversation from Python

I'm doing a simple chatbot with watson. I have a python script. Assume script is this for simplicity:
x=5
x
And in watson i want to return :
result is 5
However, I'm not sure, how to interact with python. My research showed that it is something related to NodeJS and JSON, but I couldn't find any example or tutorial that suites my requirements.
Could someone route me what course of actions should i take or any documentation?
The data between Watson Assistant and a client, an application, is exchanged as JSON-formatted data. The service itself has a REST API and you can use it from any programming language or with command-line tools. For Python, there is even a SDK.
There are some samples written in Python. I recommend my own code :). There is a tool I wrote to interact with Watson Assistant / Watson Conversation (blog entry is here). Another Python sample is what I called EgoBot (blog is here). It shows how you can even change the dialog itself from within the chatbot. Basically, you can tell the bot to learn stuff. The examples should get you started.

Does Video.io support Xamarin forms

I want to integrate a video and text chat for web (python) and mobile (xamarin). The requirement is to make customers able to communicate across platforms i.e mobile to web and vice versa.
Mobile application will be developed using Xamarin technology and the website will be in Python.
Is video supported by both platforms?
Yes it does, check out the vidyo.io Xamarin sample on github - https://github.com/Vidyo/VidyoConnector-xamarin
Note that vidyo.io SDK runs on the client end. That means it will either run on your mobile device (using native libraries or in your case using Xamarin), or in case of web sites it will use javascript. You do not need to integrate vidyo.io SDK on the server side.
On your server, what you will need, is to generate "tokens" whenever your client device or website wants to connect to a video conference.
Since you are using python you can take a look at their sample token generator at https://static.vidyo.io/4.1.16.8/utils/generateToken.py
You can learn more about tokens on their website -
https://developer.vidyo.io/documentation/4-1-16-8/getting-started#Tokens

How to receive answer from Google Assistant as a String, not as an audio stream

I am using the python libraries from the Assistant SDK for speech recognition via gRPC. I have the speech recognized and returned as a string calling the method resp.result.spoken_request_text from \googlesamples\assistant\__main__.py and I have the answer as an audio stream from the assistant API with the method resp.audio_out.audio_data also from \googlesamples\assistant\__main__.py
I would like to know if it is possible to have the answer from the service as a string as well (hoping it is available in the service definition or that it could be included), and how I could access/request the answer as string.
Thanks in advance.
Currently (Assistant SDK Developer Preview 1), there is no direct way to do this. You can probably feed the audio stream into a Speech-to-Text system, but that really starts getting silly.
Speaking to the engineers on this subject while at Google I/O, they indicated that there are some technical complications on their end to doing this, but they understand the use cases. They need to see questions like this to know that people want the feature.
Hopefully it will make it into an upcoming Developer Preview.
Update: for
google.assistant.embedded.v1alpha2
the assistant SDK includes the field supplemental_display_text
which is meant to extract the assistant response as text which aids
the user's understanding
or to be displayed on screens. Still making the text available to the developer. Goolge assistant documentation

Google Cloud Speech API real time recognition

I am developing a Python application for real-time translation. I need to recognize speech in real time: as user says something it automatically sends this piece of audio to Google Speech API and returns a text. So I want the recognized text appearing immediately while speaking.
I've found Streaming Speech Recognition but it seems that I still need to record the full speech first and then send it to the server. Also, there are no examples of how to use it in Python
Is it possible to do this with Google Speech API?
You can do it with Google Speech API.
But, it has a 1 minute content limit.
Please check the link below.
https://cloud.google.com/speech/quotas
So you have to restart every 1 minute.
and the link below is example code of microphone streaming by python.
https://cloud.google.com/speech/docs/streaming-recognize#speech-streaming-recognize-python
Check this link out:
https://github.com/Uberi/speech_recognition/blob/master/examples/microphone_recognition.py
This is an example for obtaining audio from the microphone. There are several components for the recognition process. In my experience the Sphinx Recognition lacks on accuracy. The Google Speech Recognition works very well.
Working with Google Speech API for real-time transcription is a bit cumbersome. You can use this repository for inspiration
https://github.com/saharmor/realtime-transcription
It transcribes client-side's microphone in real-time (disclaimer: I'm the author).

Categories

Resources