provide scripted test-scenarios so we can repeat tests and get comparable results
We should have more testing and testing should be reproducible and comparable etc.. Here is a basic test-sequence:
- Hans (only german) und Chad (only english) have a conversation
- two interpreters (DE->EN and EN->DE)
- settings:
Person | floor-mic | Available Language | interpeter-Mic |
---|---|---|---|
Hans | on | DE | off |
Chad | on | EN | off |
interpreter DE->EN | off | floor | EN |
interpreter EN->DE | off | floor | DE |
config of the client and specific version should be documented here
- sequence:
# | Hans says | Chad says | Chad hears | Hans hears |
---|---|---|---|---|
1 | Hallo! Wie geht es dir? | Hello! How are you? | ||
2 | Wie geht es dir? | |||
3 | Hi! I am fine thank you. | Hallo! Danke mir geht es gut. | ||
4 | I can even understand you! | Ich kann dich sogar verstehen! | ||
5 | Das war ja einfach. | That was easy. | ||
6 | Ich kann ja doch Deutsch. | Ich kann ja doch Deutsch. |
- at step 6 Hans hears the original audio, otherwise both hear the interpretation
- instead of having two interpreters, we could have one who quickly switches back and forth between the languages
I would like to document these cases in the wiki and i would like some feedback before i do. What am i missing? Would you contribute your test cases?