-
公开(公告)号:US20210204001A1
公开(公告)日:2021-07-01
申请号:US17108353
申请日:2020-12-01
Inventor: Wannan Xie , Wei Fan , Jie Li
IPC: H04N21/2187 , H04L29/08 , H04N21/433 , H04N21/6405 , G10L17/06
Abstract: Embodiments of the present disclosure provide a method for video recording, an apparatus for video recording, and a terminal device. The method can be applied to a first terminal. The first terminal can be configured to play a live video of a second terminal when the second terminal is on live. The method can include: obtaining a user identifier of a target audience logging into the first terminal; obtaining voice data of an anchor in the live video within a time period in response to detecting a first event, wherein the time period is after a current time point; and generating a video through screen recording based on the user identifier and the voice data.
-
公开(公告)号:US11405659B2
公开(公告)日:2022-08-02
申请号:US17108353
申请日:2020-12-01
Inventor: Wannan Xie , Wei Fan , Jie Li
IPC: H04N21/2187 , G10L17/06 , H04L67/306 , H04N21/433 , H04N21/6405
Abstract: Embodiments of the present disclosure provide a method for video recording, an apparatus for video recording, and a terminal device. The method can be applied to a first terminal. The first terminal can be configured to play a live video of a second terminal when the second terminal is on live. The method can include: obtaining a user identifier of a target audience logging into the first terminal; obtaining voice data of an anchor in the live video within a time period in response to detecting a first event, wherein the time period is after a current time point; and generating a video through screen recording based on the user identifier and the voice data.
-
公开(公告)号:US11482208B2
公开(公告)日:2022-10-25
申请号:US16891593
申请日:2020-06-03
Inventor: Jie Li , Xiaorui Wang , Yan Li
IPC: G10L15/02 , G06F40/166 , G06F40/242 , G10L15/187
Abstract: Disclosed are a method, device and readable storage medium for speech recognition. The method includes: determining speech features of the speech data by feature extraction on the speech data; determining syllable data corresponding to each of the speech features based on a plurality of feature extraction layers and a softmax function layer included in an acoustic model, where the acoustic model is configured to convert the speech feature into the syllable data; determining text data corresponding to the speech data based on a language model, a pronouncing dictionary and the syllable data, where the pronouncing dictionary is configured to convert the syllable data into the text data, and the language model is configured to evaluate the text data; and outputting the text data.
-
-