Update llama.cpp binaries #115
Annotations
1 error and 20 warnings
Test (macos-release)
This request was automatically failed because there were no enabled runners online to process the request for more than 1 days.
|
Test (linux-release):
LLama/Native/SafeLLamaContextHandle.cs#L203
'NativeApi.llama_eval(SafeLLamaContextHandle, int*, int, int)' is obsolete: 'use llama_decode() instead'
|
Test (linux-release):
LLama/LLamaInteractExecutor.cs#L92
Possible null reference argument for parameter 'data' in 'Task InteractiveExecutor.LoadState(ExecutorBaseState data)'.
|
Test (linux-release):
LLama/LLamaInstructExecutor.cs#L109
Possible null reference argument for parameter 'data' in 'Task InstructExecutor.LoadState(ExecutorBaseState data)'.
|
Test (linux-release):
LLama/LLamaExecutorBase.cs#L297
'LLamaContext.DeTokenize(IReadOnlyList<int>)' is obsolete: 'Use a `StreamingTokenDecoder` instead'
|
Test (linux-release):
LLama/LLamaInteractExecutor.cs#L137
'IReadOnlyListExtensions.TokensEndsWithAnyString<TTokens>(TTokens, IList<string>?, SafeLlamaModelHandle, Encoding)' is obsolete: 'Use an Antiprompt processor instead'
|
Test (linux-release):
LLama/LLamaInteractExecutor.cs#L133
This async method lacks 'await' operators and will run synchronously. Consider using the 'await' operator to await non-blocking API calls, or 'await Task.Run(...)' to do CPU-bound work on a background thread.
|
Test (linux-release):
LLama/LLamaInstructExecutor.cs#L154
'IReadOnlyListExtensions.TokensEndsWithAnyString<TTokens>(TTokens, IList<string>?, SafeLlamaModelHandle, Encoding)' is obsolete: 'Use an Antiprompt processor instead'
|
Test (linux-release):
LLama/LLamaInstructExecutor.cs#L150
This async method lacks 'await' operators and will run synchronously. Consider using the 'await' operator to await non-blocking API calls, or 'await Task.Run(...)' to do CPU-bound work on a background thread.
|
Test (linux-release):
LLama/LLamaInstructExecutor.cs#L210
Possible null reference argument for parameter 'filename' in 'void StatefulExecutorBase.SaveSessionFile(string filename)'.
|
Test (linux-release):
LLama/LLamaInteractExecutor.cs#L189
Possible null reference argument for parameter 'filename' in 'void StatefulExecutorBase.SaveSessionFile(string filename)'.
|
Test (windows-release):
LLama/Native/SafeLLamaContextHandle.cs#L203
'NativeApi.llama_eval(SafeLLamaContextHandle, int*, int, int)' is obsolete: 'use llama_decode() instead'
|
Test (windows-release):
LLama/LLamaInteractExecutor.cs#L92
Possible null reference argument for parameter 'data' in 'Task InteractiveExecutor.LoadState(ExecutorBaseState data)'.
|
Test (windows-release):
LLama/LLamaInteractExecutor.cs#L137
'IReadOnlyListExtensions.TokensEndsWithAnyString<TTokens>(TTokens, IList<string>?, SafeLlamaModelHandle, Encoding)' is obsolete: 'Use an Antiprompt processor instead'
|
Test (windows-release):
LLama/LLamaInteractExecutor.cs#L133
This async method lacks 'await' operators and will run synchronously. Consider using the 'await' operator to await non-blocking API calls, or 'await Task.Run(...)' to do CPU-bound work on a background thread.
|
Test (windows-release):
LLama/LLamaInteractExecutor.cs#L189
Possible null reference argument for parameter 'filename' in 'void StatefulExecutorBase.SaveSessionFile(string filename)'.
|
Test (windows-release):
LLama/LLamaInteractExecutor.cs#L159
This async method lacks 'await' operators and will run synchronously. Consider using the 'await' operator to await non-blocking API calls, or 'await Task.Run(...)' to do CPU-bound work on a background thread.
|
Test (windows-release):
LLama/LLamaInstructExecutor.cs#L109
Possible null reference argument for parameter 'data' in 'Task InstructExecutor.LoadState(ExecutorBaseState data)'.
|
Test (windows-release):
LLama/LLamaInstructExecutor.cs#L154
'IReadOnlyListExtensions.TokensEndsWithAnyString<TTokens>(TTokens, IList<string>?, SafeLlamaModelHandle, Encoding)' is obsolete: 'Use an Antiprompt processor instead'
|
Test (windows-release):
LLama/LLamaInstructExecutor.cs#L150
This async method lacks 'await' operators and will run synchronously. Consider using the 'await' operator to await non-blocking API calls, or 'await Task.Run(...)' to do CPU-bound work on a background thread.
|
Test (windows-release):
LLama/LLamaInstructExecutor.cs#L210
Possible null reference argument for parameter 'filename' in 'void StatefulExecutorBase.SaveSessionFile(string filename)'.
|