update readme in speechx examples,test=doc

pull/2685/head
Your Name 3 years ago
parent 4542684694
commit 0b1a45e3ae

@ -1,5 +1,16 @@
# Aishell - Deepspeech2 Streaming # Aishell - Deepspeech2 Streaming
## tutorial
## how to use in user's project
> Users want to learn how to use the example for their project ! Not just test standard result ! If users want to use these codes in their project, how to use our code ? For example, there are some audios that needs to be recognized, how to get the result . Please show it in README
> What are the requirements for audio format ?
> if there are some bins or exes builded by speechxsuch as `recognizer_main`, how to find the enterpoint source code ? how to use these program ? What are the meanings of these parameters ? What are the limits of these parameters ? Please show it in README, not just write it in run.sh .
## How to run ## How to run
``` ```

@ -1,5 +1,7 @@
# DeepSpeech2 to ONNX model # DeepSpeech2 to ONNX model
## introduce
1. convert deepspeech2 model to ONNX, using Paddle2ONNX. 1. convert deepspeech2 model to ONNX, using Paddle2ONNX.
2. check paddleinference and onnxruntime output equal. 2. check paddleinference and onnxruntime output equal.
3. optimize onnx model 3. optimize onnx model
@ -31,7 +33,18 @@ onnxruntime 1.11.0
bash run.sh --stage 0 --stop_stage 5 bash run.sh --stage 0 --stop_stage 5
``` ```
For more details please see `run.sh`. **For more details please see `run.sh`.**
> write more detail in REAME, what's meaning of different stage ?
## how to use in user's project
> Users want to learn how to use the example for their project ! Not just test standard result ! If users want to use these codes in their project, how to use our code ? For example, there are some audios that needs to be recognized, how to get the result . Please show it in README
> What are the requirements for audio format ?
> if there are some bins or exes builded by speechxsuch as `recognizer_main`, how to find the enterpoint source code ? how to use these program ? What are the meanings of these parameters ? What are the limits of these parameters ? Please show it in README, not just write it in run.sh .
## Outputs ## Outputs
The optimized onnx model is `exp/model.opt.onnx`, quanted model is `$exp/model.optset11.quant.onnx`. The optimized onnx model is `exp/model.opt.onnx`, quanted model is `$exp/model.optset11.quant.onnx`.

@ -0,0 +1,22 @@
# Deepspeech2 Websocket Demo
## introduce
## tutorial
### how to start server
> if there are some bins or exes builded by speechxsuch as `recognizer_main`, how to find the enterpoint source code ?
### how to start client
> What are the requirements for audio format ?
> What are the meanings of these parameters ?
> if there are some bins or exes builded by speechxsuch as `recognizer_main`, how to find the enterpoint source code ?
### how to write client by other language ?
> 1. websocket requirements
> 2. What are the meanings of these parameters ?
> 3. What are the limits of these parameters ?
> 4. how to write client

@ -1,5 +1,18 @@
# u2/u2pp Streaming ASR # u2/u2pp Streaming ASR
## introduce
## tutorial
### how to use in user's project
> Users want to learn how to use the example for their project ! Not just test standard result ! If users want to use these codes in their project, how to use our code ? For example, there are some audios that needs to be recognized, how to get the result . Please show it in README
> What are the requirements for audio format ?
> if there are some bins or exes builded by speechxsuch as `recognizer_main`, how to find the enterpoint source code ? how to use these program ? What are the meanings of these parameters ? What are the limits of these parameters ? Please show it in README, not just write it in run.sh .
## Testing with Aishell Test Data ## Testing with Aishell Test Data
### Download wav and model ### Download wav and model
@ -25,3 +38,7 @@
``` ```
./run.sh --stage 3 --stop_stage 3 ./run.sh --stage 3 --stop_stage 3
``` ```
### Result
> show result in README !

Loading…
Cancel
Save