Giter Club home page Giter Club logo

microsoft / projectoxford-clientsdk Goto Github PK

View Code? Open in Web Editor NEW
964.0 154.0 642.0 59.35 MB

This repo has moved :house: Visit our website for the latest SDKs & Samples

Home Page: https://www.microsoft.com/cognitive-services/en-us/sdk-sample

License: Other

Java 5.14% C# 8.76% HTML 22.35% CSS 0.29% Batchfile 0.01% PHP 0.09% JavaScript 0.93% Python 0.45% Ruby 0.05% Jupyter Notebook 57.44% Objective-C 4.07% Objective-C++ 0.17% C 0.03% ASP.NET 0.07% SCSS 0.15%

projectoxford-clientsdk's Introduction

This project has moved 🏠

We heard your feedback! This repo has been deprecated and each project has moved to a new home in a repo scoped by API and platform. We believe this is a more effective way for both you and us to make updates and get the latest SDKs & Samples.

You can now find the full list of SDKs and Samples on our website.

Happy coding!

projectoxford-clientsdk's People

Contributors

bhansen3 avatar ccasbre27 avatar chsienki avatar jpoon avatar lightfrenzy avatar microsoft-github-policy-service[bot] avatar msftgits avatar projectoxford avatar yungshinlintw avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

projectoxford-clientsdk's Issues

Emotion Analysis Example

i am trying to run this example, it works just fine for the url image, but the second one (reading an image from disk memory) give me this error

Exception in thread Thread-1:
Traceback (most recent call last):
File "C:\Anaconda2\lib\threading.py", line 801, in __bootstrap_inner
self.run()
File "C:\Anaconda2\lib\site-packages\spyderlib\widgets\externalshell\monitor.py", line 569, in run
self.update_remote_view()
File "C:\Anaconda2\lib\site-packages\spyderlib\widgets\externalshell\monitor.py", line 450, in update_remote_view
remote_view = make_remote_view(ns, settings, more_excluded_names)
File "C:\Anaconda2\lib\site-packages\spyderlib\widgets\externalshell\monitor.py", line 79, in make_remote_view
minmax=settings['minmax'])
File "C:\Anaconda2\lib\site-packages\spyderlib\widgets\dicteditorutils.py", line 225, in value_to_display
return display
UnboundLocalError: local variable 'display' referenced before assignment

Android Speech API LongDictation dutch language returns without translation

I'm using the Speech to text Android SDK with the sample application. However as the documentation says it should support the language dutch. So I changed the getDefaultLocale() to return 'nl-NL'. When the language code has changed the complete recognition stops. On the screen is visible that it did start with LongDictation recognition but the next message is *** Final n-BEST Results ***' without any result. The ShortDictation does work with the dutch language for some reason.

[Speech.JS] Error on Chrome for Android.

I get an error I am not sure how to overrun, when I use the library on Chrome for Android, as soon as the library is loaded:

Uncaught TypeError: Cannot assign to read only property 'speechSynthesis' of object '#<Window>' SpeechMain @ microsoft.speech.1.0.0.js:1493 Bing @ microsoft.speech.1.0.0.js:1496 (anonymous function) @ microsoft.speech.1.0.0.js:166

To reproduce and debug:

  1. Enable "USB Debug Mode" on your Android device from the developer settings and connect your phone to your PC;
  2. Put the library on an HTTPS site (Chrome wants HTTPS for "powerful features" such as mic access);
  3. Navigate from Android;
  4. From your desktop, navigate to chrome://inspect;
  5. Select the tab that is running your site;
  6. Refresh the page on Android;
  7. Chrome will show the error message.

start_date is mistake as end_date

When the query is i want to book a room from 3pm to 4pm today
I get today as end_date
{
"query": "i want to book a room from 3pm to 4pm today",
"intents": [
{
"intent": "builtin.intent.calendar.create_calendar_entry"
}
],
"entities": [
{
"entity": "room",
"type": "builtin.calendar.title"
},
{
"entity": "3pm",
"type": "builtin.calendar.start_time",
"resolution": {
"resolution_type": "builtin.datetime.time",
"time": "T15"
}
},
{
"entity": "4pm",
"type": "builtin.calendar.end_time",
"resolution": {
"resolution_type": "builtin.datetime.time",
"time": "T16"
}
},
{
"entity": "today",
"type": "builtin.calendar.end_date",
"resolution": {
"date": "2016-04-20",
"resolution_type": "builtin.datetime.date"
}
}
]
}

LinguisticsClient json parse error

When I run LinguisticsClient sample I got exception as follows:
Unexpected character encountered while parsing value: [. Path '[0].result', line 1, position 64.

It returns proper json but when it try deserialize it throw exception. Json looks valid.

return JsonConvert.DeserializeObject<TResponse>(responseContent, settings);
I tried to change settings but did not worked as well.

[{ "analyzerId" : "4fa79af1-f22c-408d-98bb-b7d7aeef7f04", "result" : [["NNP", "TO", "NNP", "NNP", "NNP", "."]] }, { "analyzerId" : "22a6b758-420f-4745-8a3c-46835a67c0d2", "result" : ["(TOP (INTJ (UH Welcome) (PP (TO to) (NP (NNP Microsoft) (NNP Linguistic) (NNP Analysis))) (. !)))"] }, { "analyzerId" : "08ea174b-bfdb-4e64-987e-602f85da7f72", "result" : [{ "Len" : 41, "Offset" : 0, "Tokens" : [{ "Len" : 7, "NormalizedToken" : "Welcome", "Offset" : 0, "RawToken" : "Welcome" }, { "Len" : 2, "NormalizedToken" : "to", "Offset" : 8, "RawToken" : "to" }, { "Len" : 9, "NormalizedToken" : "Microsoft", "Offset" : 11, "RawToken" : "Microsoft" }, { "Len" : 10, "NormalizedToken" : "Linguistic", "Offset" : 21, "RawToken" : "Linguistic" }, { "Len" : 8, "NormalizedToken" : "Analysis", "Offset" : 32, "RawToken" : "Analysis" }, { "Len" : 1, "NormalizedToken" : "!", "Offset" : 40, "RawToken" : "!" } ] } ] } ]

Start date is missing

When I enter book me a room today at 2pm, I get
{
"query": "book me a room today at 2pm",
"intents": [
{
"intent": "builtin.intent.calendar.create_calendar_entry"
}
],
"entities": [
{
"entity": "2pm",
"type": "builtin.calendar.start_time",
"resolution": {
"resolution_type": "builtin.datetime.time",
"time": "T14"
}
}
]
}v

SpeechClient.dll does not work with Windows Server 2012 R2 Datacenter

I have a ASP.NET deployed on Windows Server 2012. We use HTML5 Speech API to capture user's microphone. These microphone samples are sent to the server using sockets. On the server, we forward this samples to Project Oxford servers using SpeechClient.dll.

This dll works on Windows 10 and on Azure WebSite. Exception which I get is:
System.IO.FileNotFoundException: Could not load file or assembly 'SpeechClient.dll' or one of its dependencies. The specified module could not be found.

Please Help.

Domain Specific Model Call From VisionServiceRestClient Android Not Returning Celebrity Data

When I call the AnalysisInDomainResult method from VisionServiceRestClient, I get the following response:

{"metadata":{"format":"Jpeg","height":450,"width":467},"requestId":"9ad8713b-93ce-4bcb-8bbe-723a88895559"}

Notice that no domain specific result is returned for the "celebrities" category string that I pass to the method. When I call the service directly without using the provided library (passing the same image url), I do get celebrity data back. With the library, I only get metadata and the requestId.

Is there an additional step to get the celebrity data from the service in the response?

Thanks in advance!

Some errors when running the Linguistic.Sample.sln

Hi,
I found a difficulty when I tried to run the sample in LinguisticAnalysis by following 'Getting started'. I pasted my subscription key in the Program.cs (replace the "Your subscription key") and built the solution successfully, but I cannot ran the example successfully. Some errors appeared are as follows:

  1. The error poped when I opened the solution
    TF400324: Team Foundation services are not available from server http://vstfsng01:8080/tfs/cig.
    Technical information (for administrator):
    The remote name could not be resolved: 'vstfsng01'
    The active solution has been temporarily disconnected from source control because the server is unavailable. To attempt to reconnect to source control, close and then re-open the solution when the server is available. If you want to connect this solution to another server, use the Change Source Control dialog.
    The mappings for the solution could not be found.
    The mappings for the solution could not be found.
    The active solution has been temporarily disconnected from source control because the server is unavailable. To attempt to reconnect to source control, close and then re-open the solution when the server is available. If you want to connect this solution to another server, use the Change Source Control dialog.
  2. The errors after I pressed F5
    'Microsoft.ProjectOxford.Linguistics.Sample.vshost.exe' (CLR v4.0.30319: Microsoft.ProjectOxford.Linguistics.Sample.vshost.exe): Loaded 'C:\Windows\Microsoft.Net\assembly\GAC_MSIL\System.Runtime.Serialization\v4.0_4.0.0.0__b77a5c561934e089\System.Runtime.Serialization.dll'. Skipped loading symbols. Module is optimized and the debugger option 'Just My Code' is enabled.
    'Microsoft.ProjectOxford.Linguistics.Sample.vshost.exe' (CLR v4.0.30319: Microsoft.ProjectOxford.Linguistics.Sample.vshost.exe): Loaded 'C:\Windows\Microsoft.Net\assembly\GAC_MSIL\System.Xml.Linq\v4.0_4.0.0.0__b77a5c561934e089\System.Xml.Linq.dll'. Skipped loading symbols. Module is optimized and the debugger option 'Just My Code' is enabled.
    'Microsoft.ProjectOxford.Linguistics.Sample.vshost.exe' (CLR v4.0.30319: Microsoft.ProjectOxford.Linguistics.Sample.vshost.exe): Loaded 'C:\Windows\Microsoft.Net\assembly\GAC_32\System.Data\v4.0_4.0.0.0__b77a5c561934e089\System.Data.dll'. Skipped loading symbols. Module is optimized and the debugger option 'Just My Code' is enabled.
    'Microsoft.ProjectOxford.Linguistics.Sample.vshost.exe' (CLR v4.0.30319: Microsoft.ProjectOxford.Linguistics.Sample.vshost.exe): Loaded 'C:\Windows\Microsoft.Net\assembly\GAC_MSIL\System.Diagnostics.Debug\v4.0_4.0.0.0__b03f5f7f11d50a3a\System.Diagnostics.Debug.dll'. Skipped loading symbols. Module is optimized and the debugger option 'Just My Code' is enabled.
    Exception thrown: 'Newtonsoft.Json.JsonReaderException' in Newtonsoft.Json.dll
    Exception thrown: 'Newtonsoft.Json.JsonReaderException' in mscorlib.dll
    Exception thrown: 'System.AggregateException' in mscorlib.dll
    Exception thrown: 'System.Exception' in Microsoft.ProjectOxford.Linguistics.Sample.exe

Thank you very much! :)

Unclear API usage

I'm not clear on what the right sequence for using the API is.

My scenario is using the Face API for logging into an app.

The app starts with no users, the first user that's recognized should be added to a new group.
After that when a new user's face isn't recognized with sufficient confidence a new user is registered in the same group.
If an user is recognized the new face ID is added to the person's object and if there are more than 32 IDs the oldest is dropped.

(1) I saw somewhere the face IDs expire. How long do they live?
(2) what are the sequences of calls to make to the API,
(3) what is a "sufficient confidence"
(4) do I need to store images locally?

Universal app support

Installing the Emotion NuGet package into a Windows 10 universal application fails with the following error:

Installing Microsoft.ProjectOxford.Emotion 1.0.0.1.
Microsoft.ProjectOxford.Emotion 1.0.0.1 is not compatible with UAP,Version=v10.0.
Some packages are not compatible with UAP,Version=v10.0.
Microsoft.ProjectOxford.Emotion 1.0.0.1 is not compatible with UAP,Version=v10.0 (win10-arm).
Some packages are not compatible with UAP,Version=v10.0 (win10-arm).
Microsoft.ProjectOxford.Emotion 1.0.0.1 is not compatible with UAP,Version=v10.0 (win10-arm-aot).
Some packages are not compatible with UAP,Version=v10.0 (win10-arm-aot).
Microsoft.ProjectOxford.Emotion 1.0.0.1 is not compatible with UAP,Version=v10.0 (win10-x86).
Some packages are not compatible with UAP,Version=v10.0 (win10-x86).
Microsoft.ProjectOxford.Emotion 1.0.0.1 is not compatible with UAP,Version=v10.0 (win10-x86-aot).
Some packages are not compatible with UAP,Version=v10.0 (win10-x86-aot).
Microsoft.ProjectOxford.Emotion 1.0.0.1 is not compatible with UAP,Version=v10.0 (win10-x64).
Some packages are not compatible with UAP,Version=v10.0 (win10-x64).
Microsoft.ProjectOxford.Emotion 1.0.0.1 is not compatible with UAP,Version=v10.0 (win10-x64-aot).
Some packages are not compatible with UAP,Version=v10.0 (win10-x64-aot).

Any plans to support universal apps in the near future?

Thanks for all the hard work you do, it is awesome!

György

Speech Recognition hangs with wave files recorded on iOS

I am successfully using speech recognition using the data client (with and without intent) from a C# service.

However, if the wave file was recorded on iOS using the AVAudioRecorder class, none of the callbacks get called.

The difference between these wave files and others is that they have a FLLR chunk in the header, between the regular FMT and data chunks which are commonly found in wave files. If I remove the FLLR chunk, everything works fine. However, the service should probably cope with this natively.

How to call emotion api in PHP?

I try to call Emotion API via PHP program. However, it can't work. How to fix my program?

My program is showed below.

How to call JSON format in request body ?

where can I download the reuqest2.php?

Thank you


<title>無標題文件</title> $request = new Http_Request2('https://api.projectoxford.ai/emotion/v1.0/recognize');
$url = $request->getUrl();

$headers = array(
// Request headers
'Content-Type' => 'application/json',
'Ocp-Apim-Subscription-Key' => 'XXXXXX',

);

$request->setHeader($headers);

$parameters = array(
// Request parameters
'faceRectangles' => '',
);

//left,top,width,height”

$url->setQueryVariables($parameters);

$request->setMethod(HTTP_Request2::METHOD_POST);

// Request body

$request->setBody('http://chwu.weebly.com/uploads/2/9/3/6/29369481/1402061272.jpg');

try
{
$response = $request->send();
echo $response->getBody();
}
catch (HttpException $ex)
{
echo $ex;
}

?>

NPM Package for Speech.js

create a package.json and submit a package to the npm registry.

Required steps

  1. The Speech.js should not access global variables like window or navigator #61
  2. The Speech.js should be wrapped in UMD factory pattern
  3. The Speech.js should not call SpeechMain() automatically on load #61
  4. A package.json should be created and published to npm-registry

Not able to detect the emotion

While I am loading the image the then it is not able to detect any face for the purpose of detection of emotion, even it is not working for the default

Scores with Dictionary<EmotionTypesEnum, float>

Having fields instead of a collection makes it more laborious to find the best fit emotion, instead (or in addition to) the emotions' scores should be labeled with an enum and a score in a dictionary or list, so it's easy to do

emotionResults[0].Scores.OrderBy(s=>s.Value).Select(s=>s.Key).First();

If you look at your example's Log output there are 10 lines logging each value instead of one string.Join

Speech sdks cannot connect to websocket server

Both the iOS and Android app hang up at "OnHeadersAvailable@104 Auth token status: 200" after calling createMicrophoneClient method. The same code worked fine one week ago on both platforms.

The log from iOS console:

Initialize@522 ServiceUri: https://websockets.platform.bing.com/ws/speech/recognizeInitialize@589 Application Name: com.Microsoft.SpeechRecognitionServerExampleOnPreConnect@182 originating error 0x80070057 from OnPreConnect 182BeginRequest@147 originating error 0x80070057 from BeginRequest 147CloseInternal@111 CloseInternal@111 BeginRequest@149 Refreshing token /token/issueTokenAppFileToString@136 originating error 0x80004005 from AppFileToString 136InitializeRenderer@103 originating error 0x80004005 from InitializeRenderer 103Initialize@749 Initializing Audio ServicesLoadController@1567 Initializing Speech ServicesCUClient@104 Constructing CU client 0x0x7fcb79e23700RefreshState@133 Unable to get the max traffic limit from registry.RefreshState@143 Unable to get the current traffic counter from registry.ReadProperty@1041 originating error 0x8000ffff from ReadProperty 1041GetIdentityPropertyValue@235 GetIdentityPropertyValue 3GetUserAgent@255 Useragent Value iOS Assistant (iOS; 9.2;Phone;ProcessName/AppName=com.Microsoft.SpeechRecognitionServerExample;DeviceType=Near)CreateIHttpSession@98 Created WinHttp session for use by WebSocketChannel hr = (0x0)Init@126 Initializing web socket channel to endpoint https://websockets.platform.bing.com/ws/speech/recognizeInitiateHandshake@630 Initiating websocket connection. m_connection=0x0 host=websockets.platform.bing.com port=443OnHeadersAvailable@104 Auth token status: 200

The console log from Android:

02-16 18:12:47.502 12524-12524/? D/dalvikvm: Late-enabling CheckJNI
02-16 18:12:47.619 12524-12524/com.sample.test D/dalvikvm: Trying to load lib /data/data/com.sample.test/lib/libandroid_platform.so 0x41055dc0
02-16 18:12:47.650 12524-12524/com.sample.test D/dalvikvm: Added shared lib /data/data/com.sample.test/lib/libandroid_platform.so 0x41055dc0
02-16 18:12:47.650 12524-12524/com.sample.test D/dalvikvm: No JNI_OnLoad found in /data/data/com.sample.test/lib/libandroid_platform.so 0x41055dc0, skipping init
02-16 18:12:47.650 12524-12524/com.sample.test I/NLPLAT: Initialize@163 Using local IHost
02-16 18:12:47.658 12524-12524/com.sample.test I/NLPLAT: InitializeFileSystem@20 originating error 0x80070057 from InitializeFileSystem 20
02-16 18:12:47.666 12524-12524/com.sample.test I/NLPLAT: Initialize@522 ServiceUri: https://websockets.platform.bing.com/ws/speech/recognize
02-16 18:12:47.806 12524-12524/com.sample.test I/NLPLAT: OnPreConnect@182 originating error 0x80070057 from OnPreConnect 182
02-16 18:12:47.806 12524-12524/com.sample.test I/NLPLAT: BeginRequest@147 originating error 0x80070057 from BeginRequest 147
02-16 18:12:47.814 12524-12524/com.sample.test I/NLPLAT: CloseInternal@111
02-16 18:12:47.814 12524-12565/com.sample.test I/NLPLAT: Receive@491 originating error 0x8000ffff from Receive 491
02-16 18:12:47.814 12524-12565/com.sample.test I/NLPLAT: Worker@709 originating error 0x8000ffff from Worker 709
02-16 18:12:47.814 12524-12565/com.sample.test I/NLPLAT: CloseInternal@111
02-16 18:12:47.908 12524-12524/com.sample.test I/NLPLAT: BeginRequest@149 Refreshing token S
02-16 18:12:47.908 12524-12524/com.sample.test I/NLPLAT: OpenApplicationFile@52 originating error 0x80070057 from OpenApplicationFile 52
02-16 18:12:47.908 12524-12524/com.sample.test I/NLPLAT: AppFileToString@136 originating error 0x80070057 from AppFileToString 136
02-16 18:12:47.908 12524-12524/com.sample.test I/NLPLAT: InitializeRenderer@103 originating error 0x80070057 from InitializeRenderer 103
02-16 18:12:47.908 12524-12524/com.sample.test I/NLPLAT: Initialize@749 Initializing Audio Services
02-16 18:12:47.908 12524-12524/com.sample.test E/ASFDummyExtractor: isASFParserAvailable
02-16 18:12:47.908 12524-12524/com.sample.test W/ASFDummyExtractor: ASF parser is not available
02-16 18:12:47.916 12524-12524/com.sample.test W/libOpenSLES: class OutputMix interface 0 requested but unavailable MPH=43
02-16 18:12:47.924 12524-12524/com.sample.test I/NLPLAT: Load@205 Opening VAD 'raw/vcd/common/grammar/Date.0409.vad.xml'
02-16 18:12:47.924 12524-12524/com.sample.test I/NLPLAT: OpenApplicationFile@52 originating error 0x80070057 from OpenApplicationFile 52
02-16 18:12:47.924 12524-12524/com.sample.test I/NLPLAT: AppFileToString@136 originating error 0x80070057 from AppFileToString 136
02-16 18:12:47.924 12524-12524/com.sample.test I/NLPLAT: Load@207 originating error 0x80070057 from Load 207
02-16 18:12:47.924 12524-12524/com.sample.test I/NLPLAT: SetLocale@104 originating error 0x80070057 from SetLocale 104
02-16 18:12:47.924 12524-12524/com.sample.test I/NLPLAT: LoadController@1570 Initializing Speech Services
02-16 18:12:47.924 12524-12524/com.sample.test I/NLPLAT: CUClient@104 Constructing CU client 0x0x198c558
02-16 18:12:47.924 12524-12524/com.sample.test I/NLPLAT: RefreshState@133 Unable to get the max traffic limit from registry.
02-16 18:12:47.924 12524-12524/com.sample.test I/NLPLAT: RefreshState@143 Unable to get the current traffic counter from registry.
02-16 18:12:47.924 12524-12524/com.sample.test I/NLPLAT: ReadProperty@1041 originating error 0x8000ffff from ReadProperty 1041
02-16 18:12:47.924 12524-12524/com.sample.test I/NLPLAT: GetIdentityPropertyValue@235 GetIdentityPropertyValue 3
02-16 18:12:47.924 12524-12524/com.sample.test I/NLPLAT: GetUserAgent@255 Useragent Value Android Assistant (Android; Unknown;Unknown;ProcessName/AppName=Unknown;DeviceType=Near)
02-16 18:12:47.924 12524-12524/com.sample.test I/NLPLAT: CreateIHttpSession@98 Created WinHttp session for use by WebSocketChannel hr = (0x0)
02-16 18:12:47.924 12524-12524/com.sample.test I/NLPLAT: Init@126 Initializing web socket channel to endpoint https://websockets.platform.bing.com/ws/speech/recognize
02-16 18:12:47.924 12524-12524/com.sample.test I/NLPLAT: InitiateHandshake@630 Initiating websocket connection. m_connection=0x0 host=websockets.platform.bing.com port=443
02-16 18:12:47.971 12524-12571/com.sample.test I/NLPLAT: OnHeadersAvailable@104 Auth token status: 200

Getting Bad request error for enrolment - Speaker Recognition

Hi,
I've been trying to enroll a voice file for a created profile using the python API.
I was able to create a profile and list all profiles successfully. But when I try to enroll a voice (.wav) file with a simple hello world phrase with the created profile, I get the error 'ERROR:root:Error enrolling profile.' which in the trace tells 'Exception: Error enrolling profile: Bad Request'. If needed I can attach the stack trace. Can you help me getting started with this?

Luis bug: end_time is returned as start_time

{
"query": "book me a room on 4/20 from 9am to 1pm",
"intents": [
{
"intent": "builtin.intent.calendar.create_calendar_entry"
}
],
"entities": [
{
"entity": "4/20",
"type": "builtin.calendar.start_date",
"resolution": {
"date": "XXXX-04-20",
"resolution_type": "builtin.datetime.date"
}
},
{
"entity": "9am",
"type": "builtin.calendar.start_time",
"resolution": {
"resolution_type": "builtin.datetime.time",
"time": "T09"
}
},
{
"entity": "1pm",
"type": "builtin.calendar.start_time",
"resolution": {
"resolution_type": "builtin.datetime.time",
"time": "T13"
}
}
]
}

Person group ID is invalid.

the following code (vb.net vs 2015)

 Dim PersonGroupId As String = "Justme"
 Await IFaceServiceClient.CreatePersonGroupAsync(PersonGroupId, "Just_me")

generates an exception, which reads:
BadArgument
Person group ID is invalid.

Prior to this i am able to run the line:
Dim faceRects As FaceRectangle() = Await UploadAndDetectFaces(CurrentFilePath)
with results as expected.

LICENCE vs "can be used only pursuant to a non-disclosure agreement"

The LICENCE file (MIT) conflicts with the header in speech.js

// LICENSE // ------- // © 2015 Microsoft. All rights reserved. // This document is provided “as-is”. Information and views expressed in this document, including URL and other Internet Web site references, may change without notice. // Some examples depicted herein are provided for illustration only and are fictitious. No real association or connection is intended or should be inferred. // This document does not provide you with any legal rights to any intellectual property in any Microsoft product. You may copy and use this document for your internal, reference purposes. This // document is confidential and proprietary to Microsoft. It is disclosed and can be used only pursuant to a non-disclosure agreement. //----------------------------------------------------------------------

[Vision API] CancelationToken for Async calls.

Since the VisionServiceClient (C#) have no timeout we should be able to pass a CancelationToken or set a timeout for all calls to the API via the VisionServiceClient.

PS: Because of the downtime today (6.apr 2016) our integration encountered a hang.

Make it clearer where to get the client id and client secret for Speech APIs

Details are buried in different places. The Python sample says to use the subscription key for client secret, but didn't see that in the C# docs. All I see is go to the website to get the client id (which is the subscription key?) but I actually got access through the Azure Marketplace which uses primary and secondary keys. So, consider me very confused.

Could not load file or assembly 'SpeechClient, Version=0.0.0.0, Culture=neutral, PublicKeyToken=null' or one of its dependencies. An attempt was made to load a program with an incorrect format.

I am getting error while selecting any option to start recognition for speech to test project.(https://github.com/Microsoft/ProjectOxford-ClientSDK/tree/master/Speech/SpeechToText)
Error :An unhandled exception of type 'System.BadImageFormatException' occurred in PresentationCore.dll

Additional information: Could not load file or assembly 'SpeechClient, Version=0.0.0.0, Culture=neutral, PublicKeyToken=null' or one of its dependencies. An attempt was made to load a program with an incorrect format.

I have Azure account and provide primary,secondary subscriptions keys properly. Using x64.

Do we have a WebSocket API reference for continuous Speech Recognition?

Hi,

Speech.JS cannot do continuous recognition as the Partial Response is never received. I believe SpeechClient dll is connecting to Bing WebSocket servers for the continuous recognition to happen. Can't we directly connect to Bing WebSocket servers from HTML5 WebSockets? If so, then where can I find a sample code or a reference to directly connect to the Bing Speech WebSocket servers using JavaScript WebSockets?

Thanks

emotion analysis on local image with python

Can you provide a python example to show how to use the face/emotion API on local images?
I was unable to make it work. With images from internet the notebooks work well :)
Maybe you can make a python3 version of your notebooks too, just change the dict.iteritems to dict.items and add some parenthesis to the prints.

Generic exception thrown when keys have expired

I have an Azure service which makes use of multiple Project Oxford APIs, and overnight calls to each of the services started failing.

I looked through the logs, and found that the client libraries for both Vision and Face were throwing a System.Exception, with message: "Transport error".

Eventually, I found that invoking the APIs using rest directly, I got a useful message back, saying that the keys had expired. Why this happened is a separate issue, along with why I didn't get any emails to say that the keys were going to expire (they still show as active on the website). But that's all separate - this bug relates to the fact that the client libraries shouldn't just throw a generic Exception with such a vague message. The message returned by the rest API should be wrapped in the client-specific exception.

[Computer Vision Example][Python]Image and Result mismatch on local stored images

In the 'Analysis of an image stored on disk' section:
arr = np.asarray( bytearray( requests.get( urlImage ).content ), dtype=np.uint8 )
img = cv2.cvtColor( cv2.imdecode( arr, -1 ), cv2.COLOR_BGR2RGB )

The code above would fetch the online image in previous section again rather than the image stored locally. So the returned result of the local image would be rendered on the online image sample.

Using the Mac OSX SDK?

Trying to build an OSX app with the Mac OSX SDK SpeechSDK.framework, however the build fails with:

/Contents/Frameworks/SpeechSDK.framework/Versions/A: No such file or directory Command /usr/bin/codesign failed with exit code 1

Any ideas?

ARM IoT Build

Hello,

I'm trying to use the Speech SDK with the Windows IoT Platform, but unfortunately only a x64 and x86 build of the libraries are provided. Would it be possible to provide an ARM Build for the Windows IoT Platform?

AndroidSDK for Speech fails on Nexus 5 with Android 6.0.1.

I'm trying to use the speech-to-text functionality through mic and experiencing issues. Would be super grateful to get some help/pointers here.

The sample app works when built but I am unable to get it work in my own prototype. The very same prototype works on Nexus 5 with 5.1.1.

I'm using Gradle with the usual compile 'com.microsoft.projectoxford:speechrecognition:0.5.2' clause with libandroid_platform.so copied from SDK into my project's jniLibs/armeabi. Upon calling MicrophoneRecognitionClient.startMicAndRecognition() I get the following error.

01-07 22:49:23.519 30128-30128/my.app.package I/NLPLAT: Initialize@163 Using local IHost
01-07 22:49:23.521 30128-30128/my.app.package I/NLPLAT: InitializeFileSystem@20 originating error 0x80070057 from InitializeFileSystem 20
01-07 22:49:23.526 30128-30128/my.app.package I/NLPLAT: Initialize@522 ServiceUri: https://websockets.platform.bing.com/ws/speech/recognize
01-07 22:49:24.080 30128-30128/my.app.package I/NLPLAT: OnPreConnect@182 originating error 0x80070057 from OnPreConnect 182
01-07 22:49:24.080 30128-30128/my.app.package I/NLPLAT: BeginRequest@147 originating error 0x80070057 from BeginRequest 147
01-07 22:49:24.081 30128-30128/my.app.package I/NLPLAT: CloseInternal@111 
01-07 22:49:24.082 30128-1067/my.app.package I/NLPLAT: Receive@491 originating error 0x8000ffff from Receive 491
01-07 22:49:24.082 30128-1067/my.app.package I/NLPLAT: Worker@709 originating error 0x8000ffff from Worker 709
01-07 22:49:24.082 30128-1067/my.app.package I/NLPLAT: CloseInternal@111 
01-07 22:49:24.511 30128-30128/my.app.package I/NLPLAT: BeginRequest@149 Refreshing token S
01-07 22:49:24.511 30128-30128/my.app.package I/NLPLAT: OpenApplicationFile@52 originating error 0x80070057 from OpenApplicationFile 52
01-07 22:49:24.511 30128-30128/my.app.package I/NLPLAT: AppFileToString@136 originating error 0x80070057 from AppFileToString 136
01-07 22:49:24.511 30128-30128/my.app.package I/NLPLAT: InitializeRenderer@103 originating error 0x80070057 from InitializeRenderer 103
01-07 22:49:24.511 30128-30128/my.app.package I/NLPLAT: Initialize@749 Initializing Audio Services
01-07 22:49:24.512 30128-30128/my.app.package W/AudioRecord: AUDIO_INPUT_FLAG_FAST denied by client; transfer 1, track 16000 Hz, primary 48000 Hz
01-07 22:49:24.514 30128-30128/my.app.package E/AudioRecord: AudioFlinger could not create record track, status: -1
01-07 22:49:24.516 30128-30128/my.app.package E/libOpenSLES: android_audioRecorder_realize(0xae3bba00) error creating AudioRecord object; status -1
01-07 22:49:24.516 30128-30128/my.app.package W/libOpenSLES: Leaving Object::Realize (SL_RESULT_CONTENT_UNSUPPORTED)
01-07 22:49:24.516 30128-30128/my.app.package I/NLPLAT: Initialize@47 originating error 0x8000ffff from Initialize 47
01-07 22:49:24.516 30128-30128/my.app.package E/NLPLAT: native error -1910439934

It looks like the library failed to gain access to the audio source, in this case the microphone... Any ideas on how this might be resolved?

C# SDK: FaceRectangle should inherit from Rectangle

I am using the C# client SDKs available from NuGet.

When using the face detection API, you get back an array of FaceRectangles. However, the Emotion Detection API requires an array of Rectangles (from the Common namespace). Both types have the same 4 fields one would expect.

These should be consolidated - either FaceDetection should return Common Rectangles, or else FaceRectangle should inherit from Rectangle. Or, just use System.Drawing.Rectangle.

SpeechSDK for OSX

I couldn't find a github project for the iOS Speech Recognition SDK so I'm posting here.

I attempted to use the iOS Speech Recognition SDK in an OSX project and got a linker error when it was compiling. Is there a way to get the SpeechSDK.Framework to work with OSX? Or is there an OSX sdk planned for speech recognition?

C# Emotion SDK: Get ClientException without a message

I integrated the Emotion Recognition SDK in my app, using the NuGet package. However, when I call EmotionServiceClient.RecognizeAsync(), I get a ClientException, without any message.

I tried the WPF sample in this repo, with various photos, and get the same problem - ClientException without a message. The photos are <4 MB and >35x35 pixels. They work with Face Detection.

Speech.js should not call SpeechMain automatically

In environments like node.js, loading the package causes errors, like window is undefined etc.

  1. when loading the js sdk in a non-browser environment, errors should disapear.
  2. you should not access global vars (like window) but use instead root or similar.

the goal is to make the sdk more module-like #62

problems on speech sdk for ios

Hi,

I apologize for posting it here, but I couldn't really find an issue tracker for the ios SDK. So, just checking if maybe you guys can help with this, or at least provide me a link where I can find help.

I am trying to use the ios SDK, but for some reason, the flow doesn't work as it is supposed to.

I am using the MicrophoneClientWithIntent, but for some reason the method onIntentReceived is not being called as it should.

Is this a known issue? Any ideas how to workaround it?

onFinalReponseReceived is called, but there is no information about the intent at that point.

Thanks,

AndroidSDK doesn't work on Marshmallow

I have been able to run the demo on other versions of android.
When I attempt to run the demo on a 6.0 device, I get an error toast message and the following output log:

11-25 11:11:08.190 23946-25068/com.microsoft.AzureIntelligentServicesExample I/NLPLAT: ProcessMessageFailure@297 Message: 'S' has -1421792596 remaining transmit attempts.
11-25 11:11:08.190 23946-25068/com.microsoft.AzureIntelligentServicesExample I/NLPLAT: ProcessMessageFailure@302 Discarding failed message (name='S')
11-25 11:11:08.190 23946-25068/com.microsoft.AzureIntelligentServicesExample I/NLPLAT: NotifyMessageFailed@250 Reporting message failure to the observer (name='S')
11-25 11:11:08.190 23946-25068/com.microsoft.AzureIntelligentServicesExample I/NLPLAT: OnSendMessageFailed@248 Failed to 'hr', HR=8E200002, Failed to send message (name='S')
11-25 11:11:08.190 23946-25068/com.microsoft.AzureIntelligentServicesExample I/NLPLAT: ProcessTransportError@612 Transport error, hr=8E200002
11-25 11:11:08.190 23946-25068/com.microsoft.AzureIntelligentServicesExample E/NLPLAT: native error -1910505470

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.