mirror of
https://github.com/MaskRay/ccls.git
synced 2024-11-23 16:15:07 +00:00
2819 lines
88 KiB
C++
2819 lines
88 KiB
C++
// TODO: cleanup includes
|
|
#include "cache.h"
|
|
#include "code_completion.h"
|
|
#include "file_consumer.h"
|
|
#include "match.h"
|
|
#include "include_completion.h"
|
|
#include "ipc_manager.h"
|
|
#include "indexer.h"
|
|
#include "query.h"
|
|
#include "language_server_api.h"
|
|
#include "options.h"
|
|
#include "project.h"
|
|
#include "platform.h"
|
|
#include "standard_includes.h"
|
|
#include "test.h"
|
|
#include "timer.h"
|
|
#include "threaded_queue.h"
|
|
#include "working_files.h"
|
|
|
|
#include <doctest/doctest.h>
|
|
#include <rapidjson/istreamwrapper.h>
|
|
#include <rapidjson/ostreamwrapper.h>
|
|
|
|
#include <fstream>
|
|
#include <functional>
|
|
#include <iostream>
|
|
#include <iterator>
|
|
#include <string>
|
|
#include <sstream>
|
|
#include <unordered_map>
|
|
#include <thread>
|
|
#include <vector>
|
|
|
|
// TODO: provide a feature like 'https://github.com/goldsborough/clang-expand',
|
|
// ie, a fully linear view of a function with inline function calls expanded.
|
|
// We can probably use vscode decorators to achieve it.
|
|
|
|
// TODO: implement ThreadPool type which monitors CPU usage / number of work items
|
|
// per second completed and scales up/down number of running threads.
|
|
|
|
namespace {
|
|
|
|
std::vector<std::string> kEmptyArgs;
|
|
|
|
// Expected client version. We show an error if this doesn't match.
|
|
const int kExpectedClientVersion = 1;
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
#if false
|
|
std::string FormatMicroseconds(long long microseconds) {
|
|
long long milliseconds = microseconds / 1000;
|
|
long long remaining = microseconds - milliseconds;
|
|
|
|
// Only show two digits after the dot.
|
|
while (remaining >= 100)
|
|
remaining /= 10;
|
|
|
|
return std::to_string(milliseconds) + "." + std::to_string(remaining) + "ms";
|
|
}
|
|
#endif
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
// Cached completion information, so we can give fast completion results when
|
|
// the user erases a character. vscode will resend the completion request if
|
|
// that happens.
|
|
struct CodeCompleteCache {
|
|
optional<std::string> cached_path;
|
|
optional<lsPosition> cached_completion_position;
|
|
NonElidedVector<lsCompletionItem> cached_results;
|
|
NonElidedVector<lsDiagnostic> cached_diagnostics;
|
|
|
|
bool IsCacheValid(lsTextDocumentPositionParams position) const {
|
|
return cached_path == position.textDocument.uri.GetPath() &&
|
|
cached_completion_position == position.position;
|
|
}
|
|
};
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
// This function returns true if e2e timing should be displayed for the given IpcId.
|
|
bool ShouldDisplayIpcTiming(IpcId id) {
|
|
switch (id) {
|
|
case IpcId::TextDocumentPublishDiagnostics:
|
|
case IpcId::CqueryPublishInactiveRegions:
|
|
return false;
|
|
default:
|
|
return true;
|
|
}
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
bool ShouldRunIncludeCompletion(const std::string& line) {
|
|
size_t start = 0;
|
|
while (start < line.size() && isspace(line[start]))
|
|
++start;
|
|
return start < line.size() && line[start] == '#';
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
void PushBack(NonElidedVector<lsLocation>* result, optional<lsLocation> location) {
|
|
if (location)
|
|
result->push_back(*location);
|
|
}
|
|
|
|
QueryFile* FindFile(QueryDatabase* db, const std::string& filename, QueryFileId* file_id) {
|
|
auto it = db->usr_to_file.find(filename);
|
|
if (it != db->usr_to_file.end()) {
|
|
optional<QueryFile>& file = db->files[it->second.id];
|
|
if (file) {
|
|
*file_id = QueryFileId(it->second.id);
|
|
return &file.value();
|
|
}
|
|
}
|
|
|
|
std::cerr << "Unable to find file " << filename << std::endl;
|
|
*file_id = QueryFileId((size_t)-1);
|
|
return nullptr;
|
|
}
|
|
|
|
QueryFile* FindFile(QueryDatabase* db, const std::string& filename) {
|
|
// TODO: consider calling NormalizePath here. It might add too much latency though.
|
|
auto it = db->usr_to_file.find(filename);
|
|
if (it != db->usr_to_file.end()) {
|
|
optional<QueryFile>& file = db->files[it->second.id];
|
|
if (file)
|
|
return &file.value();
|
|
}
|
|
|
|
std::cerr << "Unable to find file " << filename << std::endl;
|
|
return nullptr;
|
|
}
|
|
|
|
|
|
|
|
optional<QueryLocation> GetDefinitionSpellingOfSymbol(QueryDatabase* db, const QueryTypeId& id) {
|
|
optional<QueryType>& type = db->types[id.id];
|
|
if (type)
|
|
return type->def.definition_spelling;
|
|
return nullopt;
|
|
}
|
|
|
|
optional<QueryLocation> GetDefinitionSpellingOfSymbol(QueryDatabase* db, const QueryFuncId& id) {
|
|
optional<QueryFunc>& func = db->funcs[id.id];
|
|
if (func)
|
|
return func->def.definition_spelling;
|
|
return nullopt;
|
|
}
|
|
|
|
optional<QueryLocation> GetDefinitionSpellingOfSymbol(QueryDatabase* db, const QueryVarId& id) {
|
|
optional<QueryVar>& var = db->vars[id.id];
|
|
if (var)
|
|
return var->def.definition_spelling;
|
|
return nullopt;
|
|
}
|
|
|
|
optional<QueryLocation> GetDefinitionSpellingOfSymbol(QueryDatabase* db, const SymbolIdx& symbol) {
|
|
switch (symbol.kind) {
|
|
case SymbolKind::Type: {
|
|
optional<QueryType>& type = db->types[symbol.idx];
|
|
if (type)
|
|
return type->def.definition_spelling;
|
|
break;
|
|
}
|
|
case SymbolKind::Func: {
|
|
optional<QueryFunc>& func = db->funcs[symbol.idx];
|
|
if (func)
|
|
return func->def.definition_spelling;
|
|
break;
|
|
}
|
|
case SymbolKind::Var: {
|
|
optional<QueryVar>& var = db->vars[symbol.idx];
|
|
if (var)
|
|
return var->def.definition_spelling;
|
|
break;
|
|
}
|
|
case SymbolKind::File:
|
|
case SymbolKind::Invalid: {
|
|
assert(false && "unexpected");
|
|
break;
|
|
}
|
|
}
|
|
return nullopt;
|
|
}
|
|
|
|
optional<QueryLocation> GetDefinitionExtentOfSymbol(QueryDatabase* db, const SymbolIdx& symbol) {
|
|
switch (symbol.kind) {
|
|
case SymbolKind::Type: {
|
|
optional<QueryType>& type = db->types[symbol.idx];
|
|
if (type)
|
|
return type->def.definition_extent;
|
|
break;
|
|
}
|
|
case SymbolKind::Func: {
|
|
optional<QueryFunc>& func = db->funcs[symbol.idx];
|
|
if (func)
|
|
return func->def.definition_extent;
|
|
break;
|
|
}
|
|
case SymbolKind::Var: {
|
|
optional<QueryVar>& var = db->vars[symbol.idx];
|
|
if (var)
|
|
return var->def.definition_extent;
|
|
break;
|
|
}
|
|
case SymbolKind::File: {
|
|
return QueryLocation(QueryFileId(symbol.idx), Range(Position(1, 1), Position(1, 1)));
|
|
}
|
|
case SymbolKind::Invalid: {
|
|
assert(false && "unexpected");
|
|
break;
|
|
}
|
|
}
|
|
return nullopt;
|
|
}
|
|
|
|
std::string GetHoverForSymbol(QueryDatabase* db, const SymbolIdx& symbol) {
|
|
switch (symbol.kind) {
|
|
case SymbolKind::Type: {
|
|
optional<QueryType>& type = db->types[symbol.idx];
|
|
if (type)
|
|
return type->def.detailed_name;
|
|
break;
|
|
}
|
|
case SymbolKind::Func: {
|
|
optional<QueryFunc>& func = db->funcs[symbol.idx];
|
|
if (func)
|
|
return func->def.detailed_name;
|
|
break;
|
|
}
|
|
case SymbolKind::Var: {
|
|
optional<QueryVar>& var = db->vars[symbol.idx];
|
|
if (var)
|
|
return var->def.detailed_name;
|
|
break;
|
|
}
|
|
case SymbolKind::File:
|
|
case SymbolKind::Invalid: {
|
|
assert(false && "unexpected");
|
|
break;
|
|
}
|
|
}
|
|
return "";
|
|
}
|
|
|
|
std::vector<QueryLocation> ToQueryLocation(QueryDatabase* db, const std::vector<QueryFuncRef>& refs) {
|
|
std::vector<QueryLocation> locs;
|
|
locs.reserve(refs.size());
|
|
for (const QueryFuncRef& ref : refs)
|
|
locs.push_back(ref.loc);
|
|
return locs;
|
|
}
|
|
std::vector<QueryLocation> ToQueryLocation(QueryDatabase* db, const std::vector<QueryTypeId>& ids) {
|
|
std::vector<QueryLocation> locs;
|
|
locs.reserve(ids.size());
|
|
for (const QueryTypeId& id : ids) {
|
|
optional<QueryLocation> loc = GetDefinitionSpellingOfSymbol(db, id);
|
|
if (loc)
|
|
locs.push_back(loc.value());
|
|
}
|
|
return locs;
|
|
}
|
|
std::vector<QueryLocation> ToQueryLocation(QueryDatabase* db, const std::vector<QueryFuncId>& ids) {
|
|
std::vector<QueryLocation> locs;
|
|
locs.reserve(ids.size());
|
|
for (const QueryFuncId& id : ids) {
|
|
optional<QueryLocation> loc = GetDefinitionSpellingOfSymbol(db, id);
|
|
if (loc)
|
|
locs.push_back(loc.value());
|
|
}
|
|
return locs;
|
|
}
|
|
std::vector<QueryLocation> ToQueryLocation(QueryDatabase* db, const std::vector<QueryVarId>& ids) {
|
|
std::vector<QueryLocation> locs;
|
|
locs.reserve(ids.size());
|
|
for (const QueryVarId& id : ids) {
|
|
optional<QueryLocation> loc = GetDefinitionSpellingOfSymbol(db, id);
|
|
if (loc)
|
|
locs.push_back(loc.value());
|
|
}
|
|
return locs;
|
|
}
|
|
|
|
|
|
|
|
std::vector<QueryLocation> GetUsesOfSymbol(QueryDatabase* db, const SymbolIdx& symbol) {
|
|
switch (symbol.kind) {
|
|
case SymbolKind::Type: {
|
|
optional<QueryType>& type = db->types[symbol.idx];
|
|
if (type)
|
|
return type->uses;
|
|
break;
|
|
}
|
|
case SymbolKind::Func: {
|
|
// TODO: the vector allocation could be avoided.
|
|
optional<QueryFunc>& func = db->funcs[symbol.idx];
|
|
if (func) {
|
|
std::vector<QueryLocation> result = ToQueryLocation(db, func->callers);
|
|
AddRange(&result, func->declarations);
|
|
if (func->def.definition_spelling)
|
|
result.push_back(*func->def.definition_spelling);
|
|
return result;
|
|
}
|
|
break;
|
|
}
|
|
case SymbolKind::Var: {
|
|
optional<QueryVar>& var = db->vars[symbol.idx];
|
|
if (var)
|
|
return var->uses;
|
|
break;
|
|
}
|
|
case SymbolKind::File:
|
|
case SymbolKind::Invalid: {
|
|
assert(false && "unexpected");
|
|
break;
|
|
}
|
|
}
|
|
return {};
|
|
}
|
|
|
|
std::vector<QueryLocation> GetDeclarationsOfSymbolForGotoDefinition(QueryDatabase* db, const SymbolIdx& symbol) {
|
|
switch (symbol.kind) {
|
|
case SymbolKind::Type: {
|
|
// Returning the definition spelling of a type is a hack (and is why the
|
|
// function has the postfix `ForGotoDefintion`, but it lets the user
|
|
// jump to the start of a type if clicking goto-definition on the same
|
|
// type from within the type definition.
|
|
optional<QueryType>& type = db->types[symbol.idx];
|
|
if (type) {
|
|
optional<QueryLocation> declaration = type->def.definition_spelling;
|
|
if (declaration)
|
|
return { *declaration };
|
|
}
|
|
break;
|
|
}
|
|
case SymbolKind::Func: {
|
|
optional<QueryFunc>& func = db->funcs[symbol.idx];
|
|
if (func)
|
|
return func->declarations;
|
|
break;
|
|
}
|
|
case SymbolKind::Var: {
|
|
optional<QueryVar>& var = db->vars[symbol.idx];
|
|
if (var) {
|
|
optional<QueryLocation> declaration = var->def.declaration;
|
|
if (declaration)
|
|
return { *declaration };
|
|
}
|
|
break;
|
|
}
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return {};
|
|
}
|
|
|
|
optional<QueryLocation> GetBaseDefinitionOrDeclarationSpelling(QueryDatabase* db, QueryFunc& func) {
|
|
if (!func.def.base)
|
|
return nullopt;
|
|
optional<QueryFunc>& base = db->funcs[func.def.base->id];
|
|
if (!base)
|
|
return nullopt;
|
|
|
|
auto def = base->def.definition_spelling;
|
|
if (!def && !base->declarations.empty())
|
|
def = base->declarations[0];
|
|
return def;
|
|
}
|
|
|
|
std::vector<QueryFuncRef> GetCallersForAllBaseFunctions(QueryDatabase* db, QueryFunc& root) {
|
|
std::vector<QueryFuncRef> callers;
|
|
|
|
optional<QueryFuncId> func_id = root.def.base;
|
|
while (func_id) {
|
|
optional<QueryFunc>& func = db->funcs[func_id->id];
|
|
if (!func)
|
|
break;
|
|
|
|
AddRange(&callers, func->callers);
|
|
func_id = func->def.base;
|
|
}
|
|
|
|
return callers;
|
|
}
|
|
|
|
std::vector<QueryFuncRef> GetCallersForAllDerivedFunctions(QueryDatabase* db, QueryFunc& root) {
|
|
std::vector<QueryFuncRef> callers;
|
|
|
|
std::queue<QueryFuncId> queue;
|
|
PushRange(&queue, root.derived);
|
|
|
|
while (!queue.empty()) {
|
|
optional<QueryFunc>& func = db->funcs[queue.front().id];
|
|
queue.pop();
|
|
if (!func)
|
|
continue;
|
|
|
|
PushRange(&queue, func->derived);
|
|
AddRange(&callers, func->callers);
|
|
}
|
|
|
|
return callers;
|
|
}
|
|
|
|
optional<lsRange> GetLsRange(WorkingFile* working_file, const Range& location) {
|
|
if (!working_file) {
|
|
return lsRange(
|
|
lsPosition(location.start.line - 1, location.start.column - 1),
|
|
lsPosition(location.end.line - 1, location.end.column - 1));
|
|
}
|
|
|
|
optional<int> start = working_file->GetBufferLineFromIndexLine(location.start.line);
|
|
optional<int> end = working_file->GetBufferLineFromIndexLine(location.end.line);
|
|
if (!start || !end)
|
|
return nullopt;
|
|
|
|
return lsRange(
|
|
lsPosition(*start - 1, location.start.column - 1),
|
|
lsPosition(*end - 1, location.end.column - 1));
|
|
}
|
|
|
|
lsDocumentUri GetLsDocumentUri(QueryDatabase* db, QueryFileId file_id, std::string* path) {
|
|
optional<QueryFile>& file = db->files[file_id.id];
|
|
if (file) {
|
|
*path = file->def.path;
|
|
return lsDocumentUri::FromPath(*path);
|
|
}
|
|
else {
|
|
*path = "";
|
|
return lsDocumentUri::FromPath("");
|
|
}
|
|
}
|
|
|
|
lsDocumentUri GetLsDocumentUri(QueryDatabase* db, QueryFileId file_id) {
|
|
optional<QueryFile>& file = db->files[file_id.id];
|
|
if (file) {
|
|
return lsDocumentUri::FromPath(file->def.path);
|
|
}
|
|
else {
|
|
return lsDocumentUri::FromPath("");
|
|
}
|
|
}
|
|
|
|
optional<lsLocation> GetLsLocation(QueryDatabase* db, WorkingFiles* working_files, const QueryLocation& location) {
|
|
std::string path;
|
|
lsDocumentUri uri = GetLsDocumentUri(db, location.path, &path);
|
|
optional<lsRange> range = GetLsRange(working_files->GetFileByFilename(path), location.range);
|
|
if (!range)
|
|
return nullopt;
|
|
return lsLocation(uri, *range);
|
|
}
|
|
|
|
NonElidedVector<lsLocation> GetLsLocations(QueryDatabase* db, WorkingFiles* working_files, const std::vector<QueryLocation>& locations) {
|
|
std::unordered_set<lsLocation> unique_locations;
|
|
for (const QueryLocation& query_location : locations) {
|
|
optional<lsLocation> location = GetLsLocation(db, working_files, query_location);
|
|
if (!location)
|
|
continue;
|
|
unique_locations.insert(*location);
|
|
}
|
|
|
|
NonElidedVector<lsLocation> result;
|
|
result.reserve(unique_locations.size());
|
|
result.assign(unique_locations.begin(), unique_locations.end());
|
|
return result;
|
|
}
|
|
|
|
// Returns a symbol. The symbol will have *NOT* have a location assigned.
|
|
optional<lsSymbolInformation> GetSymbolInfo(QueryDatabase* db, WorkingFiles* working_files, SymbolIdx symbol) {
|
|
switch (symbol.kind) {
|
|
case SymbolKind::File: {
|
|
optional<QueryFile>& file = db->files[symbol.idx];
|
|
if (!file)
|
|
return nullopt;
|
|
|
|
lsSymbolInformation info;
|
|
info.name = file->def.path;
|
|
info.kind = lsSymbolKind::File;
|
|
return info;
|
|
}
|
|
case SymbolKind::Type: {
|
|
optional<QueryType>& type = db->types[symbol.idx];
|
|
if (!type)
|
|
return nullopt;
|
|
|
|
lsSymbolInformation info;
|
|
info.name = type->def.short_name;
|
|
if (type->def.detailed_name != type->def.short_name)
|
|
info.containerName = type->def.detailed_name;
|
|
info.kind = lsSymbolKind::Class;
|
|
return info;
|
|
}
|
|
case SymbolKind::Func: {
|
|
optional<QueryFunc>& func = db->funcs[symbol.idx];
|
|
if (!func)
|
|
return nullopt;
|
|
|
|
lsSymbolInformation info;
|
|
info.name = func->def.short_name;
|
|
info.containerName = func->def.detailed_name;
|
|
info.kind = lsSymbolKind::Function;
|
|
|
|
if (func->def.declaring_type.has_value()) {
|
|
optional<QueryType>& container = db->types[func->def.declaring_type->id];
|
|
if (container)
|
|
info.kind = lsSymbolKind::Method;
|
|
}
|
|
|
|
return info;
|
|
}
|
|
case SymbolKind::Var: {
|
|
optional<QueryVar>& var = db->vars[symbol.idx];
|
|
if (!var)
|
|
return nullopt;
|
|
|
|
lsSymbolInformation info;
|
|
info.name += var->def.short_name;
|
|
info.containerName = var->def.detailed_name;
|
|
info.kind = lsSymbolKind::Variable;
|
|
return info;
|
|
}
|
|
case SymbolKind::Invalid: {
|
|
return nullopt;
|
|
}
|
|
};
|
|
|
|
return nullopt;
|
|
}
|
|
|
|
struct CommonCodeLensParams {
|
|
std::vector<TCodeLens>* result;
|
|
QueryDatabase* db;
|
|
WorkingFiles* working_files;
|
|
WorkingFile* working_file;
|
|
};
|
|
|
|
void AddCodeLens(
|
|
const char* singular,
|
|
const char* plural,
|
|
CommonCodeLensParams* common,
|
|
QueryLocation loc,
|
|
const std::vector<QueryLocation>& uses,
|
|
optional<QueryLocation> excluded,
|
|
bool force_display) {
|
|
TCodeLens code_lens;
|
|
optional<lsRange> range = GetLsRange(common->working_file, loc.range);
|
|
if (!range)
|
|
return;
|
|
code_lens.range = *range;
|
|
code_lens.command = lsCommand<lsCodeLensCommandArguments>();
|
|
code_lens.command->command = "cquery.showReferences";
|
|
code_lens.command->arguments.uri = GetLsDocumentUri(common->db, loc.path);
|
|
code_lens.command->arguments.position = code_lens.range.start;
|
|
|
|
// Add unique uses.
|
|
std::unordered_set<lsLocation> unique_uses;
|
|
for (const QueryLocation& use : uses) {
|
|
if (excluded == use)
|
|
continue;
|
|
optional<lsLocation> location = GetLsLocation(common->db, common->working_files, use);
|
|
if (!location)
|
|
continue;
|
|
unique_uses.insert(*location);
|
|
}
|
|
code_lens.command->arguments.locations.assign(unique_uses.begin(),
|
|
unique_uses.end());
|
|
|
|
// User visible label
|
|
size_t num_usages = unique_uses.size();
|
|
code_lens.command->title = std::to_string(num_usages) + " ";
|
|
if (num_usages == 1)
|
|
code_lens.command->title += singular;
|
|
else
|
|
code_lens.command->title += plural;
|
|
|
|
if (force_display || unique_uses.size() > 0)
|
|
common->result->push_back(code_lens);
|
|
}
|
|
|
|
lsWorkspaceEdit BuildWorkspaceEdit(QueryDatabase* db, WorkingFiles* working_files, const std::vector<QueryLocation>& locations, const std::string& new_text) {
|
|
std::unordered_map<QueryFileId, lsTextDocumentEdit> path_to_edit;
|
|
|
|
for (auto& location : locations) {
|
|
optional<lsLocation> ls_location = GetLsLocation(db, working_files, location);
|
|
if (!ls_location)
|
|
continue;
|
|
|
|
if (path_to_edit.find(location.path) == path_to_edit.end()) {
|
|
path_to_edit[location.path] = lsTextDocumentEdit();
|
|
|
|
optional<QueryFile>& file = db->files[location.path.id];
|
|
if (!file)
|
|
continue;
|
|
|
|
const std::string& path = file->def.path;
|
|
path_to_edit[location.path].textDocument.uri = lsDocumentUri::FromPath(path);
|
|
|
|
WorkingFile* working_file = working_files->GetFileByFilename(path);
|
|
if (working_file)
|
|
path_to_edit[location.path].textDocument.version = working_file->version;
|
|
}
|
|
|
|
lsTextEdit edit;
|
|
edit.range = ls_location->range;
|
|
edit.newText = new_text;
|
|
|
|
// vscode complains if we submit overlapping text edits.
|
|
auto& edits = path_to_edit[location.path].edits;
|
|
if (std::find(edits.begin(), edits.end(), edit) == edits.end())
|
|
edits.push_back(edit);
|
|
}
|
|
|
|
|
|
lsWorkspaceEdit edit;
|
|
for (const auto& changes : path_to_edit)
|
|
edit.documentChanges.push_back(changes.second);
|
|
return edit;
|
|
}
|
|
|
|
std::vector<SymbolRef> FindSymbolsAtLocation(WorkingFile* working_file, QueryFile* file, lsPosition position) {
|
|
std::vector<SymbolRef> symbols;
|
|
symbols.reserve(1);
|
|
|
|
int target_line = position.line + 1;
|
|
int target_column = position.character + 1;
|
|
if (working_file) {
|
|
optional<int> index_line = working_file->GetIndexLineFromBufferLine(target_line);
|
|
if (index_line)
|
|
target_line = *index_line;
|
|
}
|
|
|
|
for (const SymbolRef& ref : file->def.all_symbols) {
|
|
if (ref.loc.range.Contains(target_line, target_column))
|
|
symbols.push_back(ref);
|
|
}
|
|
|
|
// Order function symbols first. This makes goto definition work better when
|
|
// used on a constructor.
|
|
std::sort(symbols.begin(), symbols.end(), [](const SymbolRef& a, const SymbolRef& b) {
|
|
if (a.idx.kind != b.idx.kind && a.idx.kind == SymbolKind::Func)
|
|
return 1;
|
|
return 0;
|
|
});
|
|
|
|
return symbols;
|
|
}
|
|
|
|
|
|
|
|
void PublishInactiveLines(WorkingFile* working_file, const std::vector<Range>& inactive) {
|
|
Out_CquerySetInactiveRegion out;
|
|
out.params.uri = lsDocumentUri::FromPath(working_file->filename);
|
|
for (Range skipped : inactive) {
|
|
optional<lsRange> ls_skipped = GetLsRange(working_file, skipped);
|
|
if (ls_skipped)
|
|
out.params.inactiveRegions.push_back(*ls_skipped);
|
|
}
|
|
IpcManager::instance()->SendOutMessageToClient(IpcId::CqueryPublishInactiveRegions, out);
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
struct Index_DoIndex {
|
|
enum class Type {
|
|
// ImportOnly is used internally for loading dependency caches. The main cc
|
|
// file is loaded with ImportThenParse, which will call ImportOnly on all
|
|
// of the dependencies. The main cc will then be parsed, which will include
|
|
// updates to all dependencies.
|
|
|
|
ImportThenParse,
|
|
Parse,
|
|
Freshen,
|
|
};
|
|
|
|
Index_DoIndex(Type type, const Project::Entry& entry, optional<std::string> content, bool is_interactive)
|
|
: type(type), entry(entry), content(content), is_interactive(is_interactive) {}
|
|
|
|
// Type of index operation.
|
|
Type type;
|
|
// Project entry for file path and file arguments.
|
|
Project::Entry entry;
|
|
// File contents that should be indexed.
|
|
optional<std::string> content;
|
|
// If this index request is in response to an interactive user session, for
|
|
// example, the user saving a file they are actively editing. We report
|
|
// additional information for interactive indexes such as the IndexUpdate
|
|
// delta as well as the diagnostics.
|
|
bool is_interactive;
|
|
};
|
|
|
|
struct Index_DoIdMap {
|
|
std::unique_ptr<IndexFile> previous;
|
|
std::unique_ptr<IndexFile> current;
|
|
optional<std::string> indexed_content;
|
|
PerformanceImportFile perf;
|
|
bool is_interactive;
|
|
|
|
explicit Index_DoIdMap(std::unique_ptr<IndexFile> current,
|
|
optional<std::string> indexed_content,
|
|
PerformanceImportFile perf,
|
|
bool is_interactive)
|
|
: current(std::move(current)),
|
|
indexed_content(indexed_content),
|
|
perf(perf),
|
|
is_interactive(is_interactive) {}
|
|
|
|
explicit Index_DoIdMap(std::unique_ptr<IndexFile> previous,
|
|
std::unique_ptr<IndexFile> current,
|
|
optional<std::string> indexed_content,
|
|
PerformanceImportFile perf,
|
|
bool is_interactive)
|
|
: previous(std::move(previous)),
|
|
current(std::move(current)),
|
|
indexed_content(indexed_content),
|
|
perf(perf),
|
|
is_interactive(is_interactive) {}
|
|
};
|
|
|
|
struct Index_OnIdMapped {
|
|
std::unique_ptr<IndexFile> previous_index;
|
|
std::unique_ptr<IndexFile> current_index;
|
|
std::unique_ptr<IdMap> previous_id_map;
|
|
std::unique_ptr<IdMap> current_id_map;
|
|
optional<std::string> indexed_content;
|
|
PerformanceImportFile perf;
|
|
bool is_interactive;
|
|
|
|
Index_OnIdMapped(const optional<std::string>& indexed_content,
|
|
PerformanceImportFile perf,
|
|
bool is_interactive)
|
|
: indexed_content(indexed_content),
|
|
perf(perf),
|
|
is_interactive(is_interactive) {}
|
|
};
|
|
|
|
struct Index_OnIndexed {
|
|
IndexUpdate update;
|
|
// Map is file path to file content.
|
|
std::unordered_map<std::string, std::string> indexed_content;
|
|
PerformanceImportFile perf;
|
|
|
|
explicit Index_OnIndexed(
|
|
IndexUpdate& update,
|
|
const optional<std::string>& indexed_content,
|
|
PerformanceImportFile perf)
|
|
: update(update), perf(perf) {
|
|
if (indexed_content) {
|
|
assert(update.files_def_update.size() == 1);
|
|
this->indexed_content[update.files_def_update[0].path] = *indexed_content;
|
|
}
|
|
}
|
|
};
|
|
|
|
using Index_DoIndexQueue = ThreadedQueue<Index_DoIndex>;
|
|
using Index_DoIdMapQueue = ThreadedQueue<Index_DoIdMap>;
|
|
using Index_OnIdMappedQueue = ThreadedQueue<Index_OnIdMapped>;
|
|
using Index_OnIndexedQueue = ThreadedQueue<Index_OnIndexed>;
|
|
|
|
void RegisterMessageTypes() {
|
|
MessageRegistry::instance()->Register<Ipc_CancelRequest>();
|
|
MessageRegistry::instance()->Register<Ipc_InitializeRequest>();
|
|
MessageRegistry::instance()->Register<Ipc_InitializedNotification>();
|
|
MessageRegistry::instance()->Register<Ipc_Exit>();
|
|
MessageRegistry::instance()->Register<Ipc_TextDocumentDidOpen>();
|
|
MessageRegistry::instance()->Register<Ipc_TextDocumentDidChange>();
|
|
MessageRegistry::instance()->Register<Ipc_TextDocumentDidClose>();
|
|
MessageRegistry::instance()->Register<Ipc_TextDocumentDidSave>();
|
|
MessageRegistry::instance()->Register<Ipc_TextDocumentRename>();
|
|
MessageRegistry::instance()->Register<Ipc_TextDocumentComplete>();
|
|
MessageRegistry::instance()->Register<Ipc_TextDocumentSignatureHelp>();
|
|
MessageRegistry::instance()->Register<Ipc_TextDocumentDefinition>();
|
|
MessageRegistry::instance()->Register<Ipc_TextDocumentDocumentHighlight>();
|
|
MessageRegistry::instance()->Register<Ipc_TextDocumentHover>();
|
|
MessageRegistry::instance()->Register<Ipc_TextDocumentReferences>();
|
|
MessageRegistry::instance()->Register<Ipc_TextDocumentDocumentSymbol>();
|
|
MessageRegistry::instance()->Register<Ipc_TextDocumentDocumentLink>();
|
|
MessageRegistry::instance()->Register<Ipc_TextDocumentCodeAction>();
|
|
MessageRegistry::instance()->Register<Ipc_TextDocumentCodeLens>();
|
|
MessageRegistry::instance()->Register<Ipc_CodeLensResolve>();
|
|
MessageRegistry::instance()->Register<Ipc_WorkspaceSymbol>();
|
|
MessageRegistry::instance()->Register<Ipc_CqueryFreshenIndex>();
|
|
MessageRegistry::instance()->Register<Ipc_CqueryVars>();
|
|
MessageRegistry::instance()->Register<Ipc_CqueryCallers>();
|
|
MessageRegistry::instance()->Register<Ipc_CqueryBase>();
|
|
MessageRegistry::instance()->Register<Ipc_CqueryDerived>();
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
} // namespace
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
bool ImportCachedIndex(Config* config,
|
|
FileConsumer::SharedState* file_consumer_shared,
|
|
Index_DoIdMapQueue* queue_do_id_map,
|
|
const std::string& tu_path,
|
|
const optional<std::string>& indexed_content) {
|
|
// TODO: only load cache if command line arguments are the same.
|
|
PerformanceImportFile tu_perf;
|
|
Timer time;
|
|
|
|
std::unique_ptr<IndexFile> tu_cache = LoadCachedIndex(config, tu_path);
|
|
tu_perf.index_load_cached = time.ElapsedMicrosecondsAndReset();
|
|
if (!tu_cache)
|
|
return true;
|
|
|
|
bool needs_reparse = false;
|
|
|
|
// Import all dependencies.
|
|
for (auto& dependency_path : tu_cache->dependencies) {
|
|
//std::cerr << "- Got dependency " << dependency_path << std::endl;
|
|
PerformanceImportFile perf;
|
|
time.Reset();
|
|
std::unique_ptr<IndexFile> cache = LoadCachedIndex(config, dependency_path);
|
|
perf.index_load_cached = time.ElapsedMicrosecondsAndReset();
|
|
if (cache && GetLastModificationTime(cache->path) == cache->last_modification_time)
|
|
file_consumer_shared->Mark(cache->path);
|
|
else
|
|
needs_reparse = true;
|
|
if (cache)
|
|
queue_do_id_map->Enqueue(Index_DoIdMap(std::move(cache), nullopt, perf, false /*is_interactive*/));
|
|
}
|
|
|
|
// Import primary file.
|
|
if (GetLastModificationTime(tu_path) == tu_cache->last_modification_time)
|
|
file_consumer_shared->Mark(tu_path);
|
|
else
|
|
needs_reparse = true;
|
|
queue_do_id_map->Enqueue(Index_DoIdMap(std::move(tu_cache), indexed_content, tu_perf, false /*is_interactive*/));
|
|
|
|
return needs_reparse;
|
|
}
|
|
|
|
void ParseFile(Config* config,
|
|
WorkingFiles* working_files,
|
|
FileConsumer::SharedState* file_consumer_shared,
|
|
Index_DoIdMapQueue* queue_do_id_map,
|
|
const Project::Entry& entry,
|
|
const optional<std::string>& indexed_content,
|
|
bool is_interactive) {
|
|
|
|
std::unique_ptr<IndexFile> cache_for_args = LoadCachedIndex(config, entry.filename);
|
|
|
|
std::string tu_path = cache_for_args ? cache_for_args->import_file : entry.filename;
|
|
const std::vector<std::string>& tu_args = entry.args;
|
|
|
|
PerformanceImportFile perf;
|
|
|
|
std::vector<std::unique_ptr<IndexFile>> indexes = Parse(
|
|
config, file_consumer_shared,
|
|
tu_path, tu_args,
|
|
entry.filename, indexed_content,
|
|
&perf);
|
|
|
|
for (std::unique_ptr<IndexFile>& new_index : indexes) {
|
|
Timer time;
|
|
|
|
// Load the cached index.
|
|
std::unique_ptr<IndexFile> cached_index;
|
|
if (cache_for_args && new_index->path == cache_for_args->path)
|
|
cached_index = std::move(cache_for_args);
|
|
else
|
|
cached_index = LoadCachedIndex(config, new_index->path);
|
|
// TODO: Enable this assert when we are no longer forcibly indexing the primary file.
|
|
//assert(!cached_index || GetLastModificationTime(new_index->path) != cached_index->last_modification_time);
|
|
|
|
// Note: we are reusing the parent perf.
|
|
perf.index_load_cached = time.ElapsedMicrosecondsAndReset();
|
|
|
|
// Publish diagnostics. We should only need to publish empty diagnostics if
|
|
// |is_interactive| is true, as we may have previously published diagnostic
|
|
// for that file. If the user has diagnostics on files they are not
|
|
// editing, then they can either edit the file, in which case
|
|
// |is_interactive| will be true, or they can change the flags cquery runs
|
|
// with, in which case vscode will get restarted.
|
|
if (is_interactive || !new_index->diagnostics.empty()) {
|
|
// Emit diagnostics.
|
|
Out_TextDocumentPublishDiagnostics diag;
|
|
diag.params.uri = lsDocumentUri::FromPath(new_index->path);
|
|
diag.params.diagnostics = new_index->diagnostics;
|
|
IpcManager::instance()->SendOutMessageToClient(IpcId::TextDocumentPublishDiagnostics, diag);
|
|
|
|
// Cache diagnostics so we can show fixits.
|
|
WorkingFile* working_file = working_files->GetFileByFilename(new_index->path);
|
|
if (working_file) {
|
|
working_file->diagnostics = new_index->diagnostics;
|
|
|
|
// Publish source ranges disabled by preprocessor.
|
|
if (is_interactive) {
|
|
// TODO: We shouldn't be updating actual indexed content here, but we
|
|
// need to use the latest indexed content for the remapping.
|
|
// TODO: We should also remap diagnostics.
|
|
if (indexed_content)
|
|
working_file->SetIndexContent(*indexed_content);
|
|
|
|
PublishInactiveLines(working_file, new_index->skipped_by_preprocessor);
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
// Any any existing dependencies to |new_index| that were there before,
|
|
// because we will not reparse them if they haven't changed.
|
|
// TODO: indexer should always include dependencies. This doesn't let us remove old dependencies.
|
|
if (cached_index) {
|
|
for (auto& dep : cached_index->dependencies) {
|
|
if (std::find(new_index->dependencies.begin(), new_index->dependencies.end(), dep) == new_index->dependencies.end())
|
|
new_index->dependencies.push_back(dep);
|
|
}
|
|
}
|
|
|
|
// Forward file content, but only for the primary file.
|
|
optional<std::string> content;
|
|
if (new_index->path == entry.filename)
|
|
content = indexed_content;
|
|
|
|
// Cache the newly indexed file. This replaces the existing cache.
|
|
// TODO: Run this as another import pipeline stage.
|
|
time.Reset();
|
|
WriteToCache(config, new_index->path, *new_index, content);
|
|
perf.index_save_to_disk = time.ElapsedMicrosecondsAndReset();
|
|
|
|
// Dispatch IdMap creation request, which will happen on querydb thread.
|
|
Index_DoIdMap response(std::move(cached_index), std::move(new_index), content, perf, is_interactive);
|
|
queue_do_id_map->Enqueue(std::move(response));
|
|
}
|
|
|
|
}
|
|
|
|
bool ResetStaleFiles(Config* config,
|
|
FileConsumer::SharedState* file_consumer_shared,
|
|
const std::string& tu_path) {
|
|
Timer time;
|
|
std::unique_ptr<IndexFile> tu_cache = LoadCachedIndex(config, tu_path);
|
|
|
|
if (!tu_cache) {
|
|
std::cerr << "[indexer] Unable to load existing index from file when freshening (dependences will not be freshened)" << std::endl;
|
|
file_consumer_shared->Mark(tu_path);
|
|
return true;
|
|
}
|
|
|
|
bool needs_reparse = false;
|
|
|
|
// Check dependencies
|
|
for (auto& dependency_path : tu_cache->dependencies) {
|
|
std::unique_ptr<IndexFile> cache = LoadCachedIndex(config, dependency_path);
|
|
if (GetLastModificationTime(cache->path) != cache->last_modification_time) {
|
|
needs_reparse = true;
|
|
file_consumer_shared->Reset(cache->path);
|
|
}
|
|
}
|
|
|
|
// Check primary file
|
|
if (GetLastModificationTime(tu_path) != tu_cache->last_modification_time) {
|
|
needs_reparse = true;
|
|
file_consumer_shared->Mark(tu_path);
|
|
}
|
|
|
|
return needs_reparse;
|
|
}
|
|
|
|
bool IndexMain_DoIndex(Config* config,
|
|
FileConsumer::SharedState* file_consumer_shared,
|
|
Project* project,
|
|
WorkingFiles* working_files,
|
|
Index_DoIndexQueue* queue_do_index,
|
|
Index_DoIdMapQueue* queue_do_id_map) {
|
|
optional<Index_DoIndex> index_request = queue_do_index->TryDequeue();
|
|
if (!index_request)
|
|
return false;
|
|
|
|
Timer time;
|
|
|
|
switch (index_request->type) {
|
|
case Index_DoIndex::Type::ImportThenParse: {
|
|
// This assumes index_request->path is a cc or translation unit file (ie,
|
|
// it is in compile_commands.json).
|
|
|
|
bool needs_reparse = ImportCachedIndex(config, file_consumer_shared, queue_do_id_map, index_request->entry.filename, index_request->content);
|
|
|
|
// If the file has been updated, we need to reparse it.
|
|
if (needs_reparse) {
|
|
// Instead of parsing the file immediately, we push the request to the
|
|
// back of the queue so we will finish all of the Import requests
|
|
// before starting to run actual index jobs. This gives the user a
|
|
// partially-correct index potentially much sooner.
|
|
index_request->type = Index_DoIndex::Type::Parse;
|
|
queue_do_index->Enqueue(std::move(*index_request));
|
|
}
|
|
break;
|
|
}
|
|
|
|
case Index_DoIndex::Type::Parse: {
|
|
// index_request->path can be a cc/tu or a dependency path.
|
|
file_consumer_shared->Reset(index_request->entry.filename);
|
|
ParseFile(config, working_files, file_consumer_shared, queue_do_id_map, index_request->entry, index_request->content, index_request->is_interactive);
|
|
break;
|
|
}
|
|
|
|
case Index_DoIndex::Type::Freshen: {
|
|
// This assumes index_request->path is a cc or translation unit file (ie,
|
|
// it is in compile_commands.json).
|
|
|
|
bool needs_reparse = ResetStaleFiles(config, file_consumer_shared, index_request->entry.filename);
|
|
if (needs_reparse)
|
|
ParseFile(config, working_files, file_consumer_shared, queue_do_id_map, index_request->entry, index_request->content, index_request->is_interactive);
|
|
break;
|
|
}
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
bool IndexMain_DoCreateIndexUpdate(
|
|
Index_OnIdMappedQueue* queue_on_id_mapped,
|
|
Index_OnIndexedQueue* queue_on_indexed) {
|
|
optional<Index_OnIdMapped> response = queue_on_id_mapped->TryDequeue();
|
|
if (!response)
|
|
return false;
|
|
|
|
Timer time;
|
|
IndexUpdate update = IndexUpdate::CreateDelta(response->previous_id_map.get(), response->current_id_map.get(),
|
|
response->previous_index.get(), response->current_index.get());
|
|
response->perf.index_make_delta = time.ElapsedMicrosecondsAndReset();
|
|
|
|
#if false
|
|
#define PRINT_SECTION(name) \
|
|
if (response->perf.name) {\
|
|
total += response->perf.name; \
|
|
long long milliseconds = response->perf.name / 1000; \
|
|
long long remaining = response->perf.name - milliseconds; \
|
|
output << " " << #name << ": " << FormatMicroseconds(response->perf.name); \
|
|
}
|
|
std::stringstream output;
|
|
long long total = 0;
|
|
output << "[perf]";
|
|
PRINT_SECTION(index_parse);
|
|
PRINT_SECTION(index_build);
|
|
PRINT_SECTION(index_save_to_disk);
|
|
PRINT_SECTION(index_load_cached);
|
|
PRINT_SECTION(querydb_id_map);
|
|
PRINT_SECTION(index_make_delta);
|
|
output << "\n total: " << FormatMicroseconds(total);
|
|
output << " path: " << response->current_index->path;
|
|
output << std::endl;
|
|
std::cerr << output.rdbuf();
|
|
#undef PRINT_SECTION
|
|
|
|
if (response->is_interactive)
|
|
std::cerr << "Applying IndexUpdate" << std::endl << update.ToString() << std::endl;
|
|
#endif
|
|
|
|
Index_OnIndexed reply(update, response->indexed_content, response->perf);
|
|
queue_on_indexed->Enqueue(std::move(reply));
|
|
|
|
return true;
|
|
}
|
|
|
|
bool IndexMergeIndexUpdates(Index_OnIndexedQueue* queue_on_indexed) {
|
|
optional<Index_OnIndexed> root = queue_on_indexed->TryDequeue();
|
|
if (!root)
|
|
return false;
|
|
|
|
bool did_merge = false;
|
|
while (true) {
|
|
optional<Index_OnIndexed> to_join = queue_on_indexed->TryDequeue();
|
|
if (!to_join) {
|
|
queue_on_indexed->Enqueue(std::move(*root));
|
|
return did_merge;
|
|
}
|
|
|
|
did_merge = true;
|
|
//Timer time;
|
|
root->update.Merge(to_join->update);
|
|
for (auto&& entry : to_join->indexed_content)
|
|
root->indexed_content.emplace(entry);
|
|
//time.ResetAndPrint("[indexer] Joining two querydb updates");
|
|
}
|
|
}
|
|
|
|
void IndexMain(
|
|
Config* config,
|
|
FileConsumer::SharedState* file_consumer_shared,
|
|
Project* project,
|
|
WorkingFiles* working_files,
|
|
MultiQueueWaiter* waiter,
|
|
Index_DoIndexQueue* queue_do_index,
|
|
Index_DoIdMapQueue* queue_do_id_map,
|
|
Index_OnIdMappedQueue* queue_on_id_mapped,
|
|
Index_OnIndexedQueue* queue_on_indexed) {
|
|
|
|
SetCurrentThreadName("indexer");
|
|
while (true) {
|
|
// TODO: process all off IndexMain_DoIndex before calling IndexMain_DoCreateIndexUpdate for
|
|
// better icache behavior. We need to have some threads spinning on both though
|
|
// otherwise memory usage will get bad.
|
|
|
|
// We need to make sure to run both IndexMain_DoIndex and
|
|
// IndexMain_DoCreateIndexUpdate so we don't starve querydb from doing any
|
|
// work. Running both also lets the user query the partially constructed
|
|
// index.
|
|
bool did_index = IndexMain_DoIndex(config, file_consumer_shared, project, working_files, queue_do_index, queue_do_id_map);
|
|
bool did_create_update = IndexMain_DoCreateIndexUpdate(queue_on_id_mapped, queue_on_indexed);
|
|
bool did_merge = false;
|
|
|
|
// Nothing to index and no index updates to create, so join some already
|
|
// created index updates to reduce work on querydb thread.
|
|
if (!did_index && !did_create_update)
|
|
did_merge = IndexMergeIndexUpdates(queue_on_indexed);
|
|
|
|
// We didn't do any work, so wait for a notification.
|
|
if (!did_index && !did_create_update && !did_merge)
|
|
waiter->Wait({
|
|
queue_do_index,
|
|
queue_on_id_mapped,
|
|
queue_on_indexed
|
|
});
|
|
}
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
bool QueryDbMainLoop(
|
|
Config* config,
|
|
QueryDatabase* db,
|
|
MultiQueueWaiter* waiter,
|
|
Index_DoIndexQueue* queue_do_index,
|
|
Index_DoIdMapQueue* queue_do_id_map,
|
|
Index_OnIdMappedQueue* queue_on_id_mapped,
|
|
Index_OnIndexedQueue* queue_on_indexed,
|
|
Project* project,
|
|
FileConsumer::SharedState* file_consumer_shared,
|
|
WorkingFiles* working_files,
|
|
CompletionManager* completion_manager,
|
|
IncludeCompletion* include_completion,
|
|
CodeCompleteCache* code_complete_cache,
|
|
CodeCompleteCache* signature_cache) {
|
|
IpcManager* ipc = IpcManager::instance();
|
|
|
|
bool did_work = false;
|
|
|
|
std::vector<std::unique_ptr<BaseIpcMessage>> messages = ipc->GetMessages(IpcManager::Destination::Server);
|
|
for (auto& message : messages) {
|
|
did_work = true;
|
|
//std::cerr << "[querydb] Processing message " << IpcIdToString(message->method_id) << std::endl;
|
|
|
|
switch (message->method_id) {
|
|
case IpcId::Initialize: {
|
|
auto request = static_cast<Ipc_InitializeRequest*>(message.get());
|
|
|
|
// Log initialization parameters.
|
|
rapidjson::StringBuffer output;
|
|
Writer writer(output);
|
|
Reflect(writer, request->params.initializationOptions);
|
|
std::cerr << output.GetString() << std::endl;
|
|
|
|
if (request->params.rootUri) {
|
|
std::string project_path = request->params.rootUri->GetPath();
|
|
std::cerr << "[querydb] Initialize in directory " << project_path
|
|
<< " with uri " << request->params.rootUri->raw_uri
|
|
<< std::endl;
|
|
|
|
if (!request->params.initializationOptions) {
|
|
std::cerr << "Initialization parameters (particularily cacheDirectory) are required" << std::endl;
|
|
exit(1);
|
|
}
|
|
|
|
*config = *request->params.initializationOptions;
|
|
|
|
// Check client version.
|
|
if (config->clientVersion != kExpectedClientVersion) {
|
|
Out_ShowLogMessage out;
|
|
out.display_type = Out_ShowLogMessage::DisplayType::Show;
|
|
out.params.type = lsMessageType::Error;
|
|
out.params.message = "cquery client (v" + std::to_string(config->clientVersion) + ") and server (v" + std::to_string(kExpectedClientVersion) + ") version mismatch. Please update ";
|
|
if (config->clientVersion > kExpectedClientVersion)
|
|
out.params.message += "the cquery binary.";
|
|
else
|
|
out.params.message += "your extension client (VSIX file).";
|
|
out.Write(std::cout);
|
|
}
|
|
|
|
// Make sure cache directory is valid.
|
|
if (config->cacheDirectory.empty()) {
|
|
std::cerr << "[fatal] No cache directory" << std::endl;
|
|
exit(1);
|
|
}
|
|
config->cacheDirectory = NormalizePath(config->cacheDirectory);
|
|
EnsureEndsInSlash(config->cacheDirectory);
|
|
MakeDirectoryRecursive(config->cacheDirectory);
|
|
|
|
// Set project root.
|
|
config->projectRoot = NormalizePath(request->params.rootUri->GetPath());
|
|
EnsureEndsInSlash(config->projectRoot);
|
|
|
|
// Start indexer threads.
|
|
int indexer_count = std::max<int>(std::thread::hardware_concurrency(), 2) - 1;
|
|
if (config->indexerCount > 0)
|
|
indexer_count = config->indexerCount;
|
|
std::cerr << "[querydb] Starting " << indexer_count << " indexers" << std::endl;
|
|
for (int i = 0; i < indexer_count; ++i) {
|
|
new std::thread([&]() {
|
|
IndexMain(config, file_consumer_shared, project, working_files, waiter, queue_do_index, queue_do_id_map, queue_on_id_mapped, queue_on_indexed);
|
|
});
|
|
}
|
|
|
|
Timer time;
|
|
|
|
// Open up / load the project.
|
|
project->Load(config->extraClangArguments, project_path);
|
|
time.ResetAndPrint("[perf] Loaded compilation entries (" + std::to_string(project->entries.size()) + " files)");
|
|
|
|
// Start scanning include directories before dispatching project files, because that takes a long time.
|
|
include_completion->Rescan();
|
|
|
|
time.Reset();
|
|
project->ForAllFilteredFiles(config, [&](int i, const Project::Entry& entry) {
|
|
//std::cerr << "[" << i << "/" << (project->entries.size() - 1)
|
|
// << "] Dispatching index request for file " << entry.filename
|
|
// << std::endl;
|
|
queue_do_index->Enqueue(Index_DoIndex(Index_DoIndex::Type::ImportThenParse, entry, nullopt, false /*is_interactive*/));
|
|
});
|
|
time.ResetAndPrint("[perf] Dispatched initial index requests");
|
|
}
|
|
|
|
// TODO: query request->params.capabilities.textDocument and support only things
|
|
// the client supports.
|
|
|
|
auto response = Out_InitializeResponse();
|
|
response.id = request->id;
|
|
|
|
//response.result.capabilities.textDocumentSync = lsTextDocumentSyncOptions();
|
|
//response.result.capabilities.textDocumentSync->openClose = true;
|
|
//response.result.capabilities.textDocumentSync->change = lsTextDocumentSyncKind::Full;
|
|
//response.result.capabilities.textDocumentSync->willSave = true;
|
|
//response.result.capabilities.textDocumentSync->willSaveWaitUntil = true;
|
|
response.result.capabilities.textDocumentSync = lsTextDocumentSyncKind::Incremental;
|
|
|
|
response.result.capabilities.renameProvider = true;
|
|
|
|
response.result.capabilities.completionProvider = lsCompletionOptions();
|
|
response.result.capabilities.completionProvider->resolveProvider = false;
|
|
// vscode doesn't support trigger character sequences, so we use ':' for '::' and '>' for '->'.
|
|
// See https://github.com/Microsoft/language-server-protocol/issues/138.
|
|
response.result.capabilities.completionProvider->triggerCharacters = { ".", ":", ">", "#" };
|
|
|
|
response.result.capabilities.signatureHelpProvider = lsSignatureHelpOptions();
|
|
// NOTE: If updating signature help tokens make sure to also update
|
|
// WorkingFile::FindClosestCallNameInBuffer.
|
|
response.result.capabilities.signatureHelpProvider->triggerCharacters = { "(", "," };
|
|
|
|
response.result.capabilities.codeLensProvider = lsCodeLensOptions();
|
|
response.result.capabilities.codeLensProvider->resolveProvider = false;
|
|
|
|
response.result.capabilities.definitionProvider = true;
|
|
response.result.capabilities.documentHighlightProvider = true;
|
|
response.result.capabilities.hoverProvider = true;
|
|
response.result.capabilities.referencesProvider = true;
|
|
|
|
response.result.capabilities.codeActionProvider = true;
|
|
|
|
response.result.capabilities.documentSymbolProvider = true;
|
|
response.result.capabilities.workspaceSymbolProvider = true;
|
|
|
|
response.result.capabilities.documentLinkProvider = lsDocumentLinkOptions();
|
|
response.result.capabilities.documentLinkProvider->resolveProvider = false;
|
|
|
|
ipc->SendOutMessageToClient(IpcId::Initialize, response);
|
|
break;
|
|
}
|
|
|
|
case IpcId::Exit: {
|
|
exit(0);
|
|
break;
|
|
}
|
|
|
|
case IpcId::CqueryFreshenIndex: {
|
|
std::cerr << "Freshening " << project->entries.size() << " files" << std::endl;
|
|
project->ForAllFilteredFiles(config, [&](int i, const Project::Entry& entry) {
|
|
std::cerr << "[" << i << "/" << (project->entries.size() - 1)
|
|
<< "] Dispatching index request for file " << entry.filename
|
|
<< std::endl;
|
|
queue_do_index->Enqueue(Index_DoIndex(Index_DoIndex::Type::Freshen, entry, nullopt, false /*is_interactive*/));
|
|
});
|
|
break;
|
|
}
|
|
|
|
case IpcId::CqueryVars: {
|
|
auto msg = static_cast<Ipc_CqueryVars*>(message.get());
|
|
|
|
QueryFile* file = FindFile(db, msg->params.textDocument.uri.GetPath());
|
|
if (!file) {
|
|
std::cerr << "Unable to find file " << msg->params.textDocument.uri.GetPath() << std::endl;
|
|
break;
|
|
}
|
|
WorkingFile* working_file = working_files->GetFileByFilename(file->def.path);
|
|
|
|
Out_LocationList response;
|
|
response.id = msg->id;
|
|
for (const SymbolRef& ref : FindSymbolsAtLocation(working_file, file, msg->params.position)) {
|
|
if (ref.idx.kind == SymbolKind::Type) {
|
|
optional<QueryType>& type = db->types[ref.idx.idx];
|
|
if (!type) continue;
|
|
std::vector<QueryLocation> locations = ToQueryLocation(db, type->instances);
|
|
response.result = GetLsLocations(db, working_files, locations);
|
|
}
|
|
}
|
|
ipc->SendOutMessageToClient(IpcId::TextDocumentReferences, response);
|
|
break;
|
|
}
|
|
|
|
case IpcId::CqueryCallers: {
|
|
auto msg = static_cast<Ipc_CqueryCallers*>(message.get());
|
|
|
|
QueryFile* file = FindFile(db, msg->params.textDocument.uri.GetPath());
|
|
if (!file) {
|
|
std::cerr << "Unable to find file " << msg->params.textDocument.uri.GetPath() << std::endl;
|
|
break;
|
|
}
|
|
WorkingFile* working_file = working_files->GetFileByFilename(file->def.path);
|
|
|
|
Out_LocationList response;
|
|
response.id = msg->id;
|
|
for (const SymbolRef& ref : FindSymbolsAtLocation(working_file, file, msg->params.position)) {
|
|
if (ref.idx.kind == SymbolKind::Func) {
|
|
optional<QueryFunc>& func = db->funcs[ref.idx.idx];
|
|
if (!func) continue;
|
|
std::vector<QueryLocation> locations = ToQueryLocation(db, func->callers);
|
|
response.result = GetLsLocations(db, working_files, locations);
|
|
}
|
|
}
|
|
ipc->SendOutMessageToClient(IpcId::TextDocumentReferences, response);
|
|
break;
|
|
}
|
|
|
|
case IpcId::CqueryBase: {
|
|
auto msg = static_cast<Ipc_CqueryBase*>(message.get());
|
|
|
|
QueryFile* file = FindFile(db, msg->params.textDocument.uri.GetPath());
|
|
if (!file) {
|
|
std::cerr << "Unable to find file " << msg->params.textDocument.uri.GetPath() << std::endl;
|
|
break;
|
|
}
|
|
WorkingFile* working_file = working_files->GetFileByFilename(file->def.path);
|
|
|
|
Out_LocationList response;
|
|
response.id = msg->id;
|
|
for (const SymbolRef& ref : FindSymbolsAtLocation(working_file, file, msg->params.position)) {
|
|
if (ref.idx.kind == SymbolKind::Type) {
|
|
optional<QueryType>& type = db->types[ref.idx.idx];
|
|
if (!type) continue;
|
|
std::vector<QueryLocation> locations = ToQueryLocation(db, type->def.parents);
|
|
response.result = GetLsLocations(db, working_files, locations);
|
|
}
|
|
else if (ref.idx.kind == SymbolKind::Func) {
|
|
optional<QueryFunc>& func = db->funcs[ref.idx.idx];
|
|
if (!func) continue;
|
|
optional<QueryLocation> location = GetBaseDefinitionOrDeclarationSpelling(db, *func);
|
|
if (!location) continue;
|
|
optional<lsLocation> ls_loc = GetLsLocation(db, working_files, *location);
|
|
if (!ls_loc) continue;
|
|
response.result.push_back(*ls_loc);
|
|
}
|
|
}
|
|
ipc->SendOutMessageToClient(IpcId::TextDocumentReferences, response);
|
|
break;
|
|
}
|
|
|
|
case IpcId::CqueryDerived: {
|
|
auto msg = static_cast<Ipc_CqueryDerived*>(message.get());
|
|
|
|
QueryFile* file = FindFile(db, msg->params.textDocument.uri.GetPath());
|
|
if (!file) {
|
|
std::cerr << "Unable to find file " << msg->params.textDocument.uri.GetPath() << std::endl;
|
|
break;
|
|
}
|
|
WorkingFile* working_file = working_files->GetFileByFilename(file->def.path);
|
|
|
|
Out_LocationList response;
|
|
response.id = msg->id;
|
|
for (const SymbolRef& ref : FindSymbolsAtLocation(working_file, file, msg->params.position)) {
|
|
if (ref.idx.kind == SymbolKind::Type) {
|
|
optional<QueryType>& type = db->types[ref.idx.idx];
|
|
if (!type) continue;
|
|
std::vector<QueryLocation> locations = ToQueryLocation(db, type->derived);
|
|
response.result = GetLsLocations(db, working_files, locations);
|
|
}
|
|
else if (ref.idx.kind == SymbolKind::Func) {
|
|
optional<QueryFunc>& func = db->funcs[ref.idx.idx];
|
|
if (!func) continue;
|
|
std::vector<QueryLocation> locations = ToQueryLocation(db, func->derived);
|
|
response.result = GetLsLocations(db, working_files, locations);
|
|
}
|
|
}
|
|
ipc->SendOutMessageToClient(IpcId::TextDocumentReferences, response);
|
|
break;
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
case IpcId::TextDocumentDidOpen: {
|
|
// NOTE: This function blocks code lens. If it starts taking a long time
|
|
// we will need to find a way to unblock the code lens request.
|
|
|
|
Timer time;
|
|
auto msg = static_cast<Ipc_TextDocumentDidOpen*>(message.get());
|
|
WorkingFile* working_file = working_files->OnOpen(msg->params);
|
|
optional<std::string> cached_file_contents = LoadCachedFileContents(config, msg->params.textDocument.uri.GetPath());
|
|
if (cached_file_contents)
|
|
working_file->SetIndexContent(*cached_file_contents);
|
|
else
|
|
working_file->SetIndexContent(working_file->buffer_content);
|
|
|
|
std::unique_ptr<IndexFile> cache = LoadCachedIndex(config, msg->params.textDocument.uri.GetPath());
|
|
if (cache && !cache->skipped_by_preprocessor.empty())
|
|
PublishInactiveLines(working_file, cache->skipped_by_preprocessor);
|
|
|
|
time.ResetAndPrint("[querydb] Loading cached index file for DidOpen (blocks CodeLens)");
|
|
|
|
include_completion->AddFile(working_file->filename);
|
|
|
|
break;
|
|
}
|
|
|
|
case IpcId::TextDocumentDidChange: {
|
|
auto msg = static_cast<Ipc_TextDocumentDidChange*>(message.get());
|
|
working_files->OnChange(msg->params);
|
|
break;
|
|
}
|
|
|
|
case IpcId::TextDocumentDidClose: {
|
|
auto msg = static_cast<Ipc_TextDocumentDidClose*>(message.get());
|
|
working_files->OnClose(msg->params);
|
|
break;
|
|
}
|
|
|
|
case IpcId::TextDocumentDidSave: {
|
|
auto msg = static_cast<Ipc_TextDocumentDidSave*>(message.get());
|
|
|
|
std::string path = msg->params.textDocument.uri.GetPath();
|
|
// Send out an index request, and copy the current buffer state so we
|
|
// can update the cached index contents when the index is done.
|
|
//
|
|
// We also do not index if there is already an index request.
|
|
//
|
|
// TODO: Cancel outgoing index request. Might be tricky to make
|
|
// efficient since we have to cancel.
|
|
// - we could have an |atomic<int> active_cancellations| variable
|
|
// that all of the indexers check before accepting an index. if
|
|
// zero we don't slow down fast-path. if non-zero we acquire
|
|
// mutex and check to see if we should skip the current request.
|
|
// if so, ignore that index response.
|
|
WorkingFile* working_file = working_files->GetFileByFilename(path);
|
|
if (working_file) {
|
|
// Only do a delta update (Type::Parse) if we've already imported the
|
|
// file. If the user saves a file not loaded by the project we don't
|
|
// want the initial import to be a delta-update.
|
|
Index_DoIndex::Type index_type = Index_DoIndex::Type::Parse;
|
|
QueryFile* file = FindFile(db, path);
|
|
if (!file)
|
|
index_type = Index_DoIndex::Type::ImportThenParse;
|
|
|
|
queue_do_index->PriorityEnqueue(Index_DoIndex(index_type, project->FindCompilationEntryForFile(path), working_file->buffer_content, true /*is_interactive*/));
|
|
}
|
|
completion_manager->UpdateActiveSession(path);
|
|
|
|
break;
|
|
}
|
|
|
|
case IpcId::TextDocumentRename: {
|
|
auto msg = static_cast<Ipc_TextDocumentRename*>(message.get());
|
|
|
|
QueryFileId file_id;
|
|
QueryFile* file = FindFile(db, msg->params.textDocument.uri.GetPath(), &file_id);
|
|
if (!file) {
|
|
std::cerr << "Unable to find file " << msg->params.textDocument.uri.GetPath() << std::endl;
|
|
break;
|
|
}
|
|
WorkingFile* working_file = working_files->GetFileByFilename(file->def.path);
|
|
|
|
Out_TextDocumentRename response;
|
|
response.id = msg->id;
|
|
|
|
for (const SymbolRef& ref : FindSymbolsAtLocation(working_file, file, msg->params.position)) {
|
|
// Found symbol. Return references to rename.
|
|
std::vector<QueryLocation> uses = GetUsesOfSymbol(db, ref.idx);
|
|
response.result = BuildWorkspaceEdit(db, working_files, uses, msg->params.newName);
|
|
break;
|
|
}
|
|
|
|
response.Write(std::cerr);
|
|
ipc->SendOutMessageToClient(IpcId::TextDocumentRename, response);
|
|
break;
|
|
}
|
|
|
|
case IpcId::TextDocumentCompletion: {
|
|
auto msg = static_cast<Ipc_TextDocumentComplete*>(message.get());
|
|
lsTextDocumentPositionParams& params = msg->params;
|
|
|
|
WorkingFile* file = working_files->GetFileByFilename(params.textDocument.uri.GetPath());
|
|
|
|
// TODO: We should scan include directories to add any missing paths
|
|
|
|
std::string buffer_line = file->all_buffer_lines[params.position.line];
|
|
|
|
if (ShouldRunIncludeCompletion(buffer_line)) {
|
|
Out_TextDocumentComplete complete_response;
|
|
complete_response.id = msg->id;
|
|
complete_response.result.isIncomplete = false;
|
|
|
|
{
|
|
std::unique_lock<std::mutex> lock(include_completion->completion_items_mutex, std::defer_lock);
|
|
if (include_completion->is_scanning)
|
|
lock.lock();
|
|
complete_response.result.items.assign(
|
|
include_completion->completion_items.begin(),
|
|
include_completion->completion_items.end());
|
|
if (lock)
|
|
lock.unlock();
|
|
|
|
// Update textEdit params.
|
|
for (lsCompletionItem& item : complete_response.result.items) {
|
|
item.textEdit->range.start.line = params.position.line;
|
|
item.textEdit->range.start.character = 0;
|
|
item.textEdit->range.end.line = params.position.line;
|
|
item.textEdit->range.end.character = (int)buffer_line.size();
|
|
}
|
|
}
|
|
|
|
std::cerr << "[complete] Returning " << complete_response.result.items.size() << " include completions" << std::endl;
|
|
ipc->SendOutMessageToClient(IpcId::TextDocumentCompletion, complete_response);
|
|
}
|
|
else {
|
|
if (file)
|
|
params.position = file->FindStableCompletionSource(params.position);
|
|
|
|
CompletionManager::OnComplete callback = std::bind([working_files, code_complete_cache](BaseIpcMessage* message, NonElidedVector<lsCompletionItem> results, NonElidedVector<lsDiagnostic> diagnostics) {
|
|
auto msg = static_cast<Ipc_TextDocumentComplete*>(message);
|
|
auto ipc = IpcManager::instance();
|
|
|
|
Out_TextDocumentComplete complete_response;
|
|
complete_response.id = msg->id;
|
|
complete_response.result.isIncomplete = false;
|
|
complete_response.result.items = results;
|
|
|
|
// Emit completion results.
|
|
ipc->SendOutMessageToClient(IpcId::TextDocumentCompletion, complete_response);
|
|
|
|
// Emit diagnostics.
|
|
Out_TextDocumentPublishDiagnostics diagnostic_response;
|
|
diagnostic_response.params.uri = msg->params.textDocument.uri;
|
|
diagnostic_response.params.diagnostics = diagnostics;
|
|
ipc->SendOutMessageToClient(IpcId::TextDocumentPublishDiagnostics, diagnostic_response);
|
|
|
|
// Cache diagnostics so we can show fixits.
|
|
WorkingFile* working_file = working_files->GetFileByFilename(msg->params.textDocument.uri.GetPath());
|
|
if (working_file)
|
|
working_file->diagnostics = diagnostics;
|
|
|
|
// Cache completion results so if the user types backspace we can respond faster.
|
|
code_complete_cache->cached_path = msg->params.textDocument.uri.GetPath();
|
|
code_complete_cache->cached_completion_position = msg->params.position;
|
|
code_complete_cache->cached_results = results;
|
|
code_complete_cache->cached_diagnostics = diagnostics;
|
|
|
|
delete message;
|
|
}, message.release(), std::placeholders::_1, std::placeholders::_2);
|
|
|
|
if (code_complete_cache->IsCacheValid(params)) {
|
|
std::cerr << "[complete] Using cached completion results at " << params.position.ToString() << std::endl;
|
|
callback(code_complete_cache->cached_results, code_complete_cache->cached_diagnostics);
|
|
}
|
|
else {
|
|
completion_manager->CodeComplete(params, std::move(callback));
|
|
}
|
|
}
|
|
|
|
break;
|
|
}
|
|
|
|
case IpcId::TextDocumentSignatureHelp: {
|
|
auto msg = static_cast<Ipc_TextDocumentSignatureHelp*>(message.get());
|
|
lsTextDocumentPositionParams& params = msg->params;
|
|
WorkingFile* file = working_files->GetFileByFilename(params.textDocument.uri.GetPath());
|
|
std::string search;
|
|
int active_param = 0;
|
|
if (file) {
|
|
lsPosition completion_position;
|
|
search = file->FindClosestCallNameInBuffer(params.position, &active_param, &completion_position);
|
|
params.position = completion_position;
|
|
}
|
|
//std::cerr << "[completion] Returning signatures for " << search << std::endl;
|
|
if (search.empty())
|
|
break;
|
|
|
|
CompletionManager::OnComplete callback = std::bind([signature_cache](BaseIpcMessage* message, std::string search, int active_param, const NonElidedVector<lsCompletionItem>& results) {
|
|
auto msg = static_cast<Ipc_TextDocumentSignatureHelp*>(message);
|
|
auto ipc = IpcManager::instance();
|
|
|
|
Out_TextDocumentSignatureHelp response;
|
|
response.id = msg->id;
|
|
|
|
for (auto& result : results) {
|
|
if (result.label != search)
|
|
continue;
|
|
|
|
lsSignatureInformation signature;
|
|
signature.label = result.detail;
|
|
for (auto& parameter : result.parameters_) {
|
|
lsParameterInformation ls_param;
|
|
ls_param.label = parameter;
|
|
signature.parameters.push_back(ls_param);
|
|
}
|
|
response.result.signatures.push_back(signature);
|
|
}
|
|
|
|
// Guess the signature the user wants based on available parameter
|
|
// count.
|
|
response.result.activeSignature = 0;
|
|
for (size_t i = 0; i < response.result.signatures.size(); ++i) {
|
|
if (active_param < response.result.signatures.size()) {
|
|
response.result.activeSignature = (int)i;
|
|
break;
|
|
}
|
|
}
|
|
|
|
// Set signature to what we parsed from the working file.
|
|
response.result.activeParameter = active_param;
|
|
|
|
response.Write(std::cerr);
|
|
std::cerr << std::endl;
|
|
|
|
Timer timer;
|
|
ipc->SendOutMessageToClient(IpcId::TextDocumentSignatureHelp, response);
|
|
timer.ResetAndPrint("[complete] Writing signature help results");
|
|
|
|
signature_cache->cached_path = msg->params.textDocument.uri.GetPath();
|
|
signature_cache->cached_completion_position = msg->params.position;
|
|
signature_cache->cached_results = results;
|
|
|
|
delete message;
|
|
}, message.release(), search, active_param, std::placeholders::_1);
|
|
|
|
if (signature_cache->IsCacheValid(params)) {
|
|
std::cerr << "[complete] Using cached completion results at " << params.position.ToString() << std::endl;
|
|
callback(signature_cache->cached_results, signature_cache->cached_diagnostics);
|
|
}
|
|
else {
|
|
completion_manager->CodeComplete(params, std::move(callback));
|
|
}
|
|
|
|
break;
|
|
}
|
|
|
|
case IpcId::TextDocumentDefinition: {
|
|
auto msg = static_cast<Ipc_TextDocumentDefinition*>(message.get());
|
|
|
|
QueryFileId file_id;
|
|
QueryFile* file = FindFile(db, msg->params.textDocument.uri.GetPath(), &file_id);
|
|
if (!file) {
|
|
std::cerr << "Unable to find file " << msg->params.textDocument.uri.GetPath() << std::endl;
|
|
break;
|
|
}
|
|
WorkingFile* working_file = working_files->GetFileByFilename(file->def.path);
|
|
|
|
Out_TextDocumentDefinition response;
|
|
response.id = msg->id;
|
|
|
|
int target_line = msg->params.position.line + 1;
|
|
int target_column = msg->params.position.character + 1;
|
|
|
|
for (const SymbolRef& ref : FindSymbolsAtLocation(working_file, file, msg->params.position)) {
|
|
// Found symbol. Return definition.
|
|
|
|
// Special cases which are handled:
|
|
// - symbol has declaration but no definition (ie, pure virtual)
|
|
// - start at spelling but end at extent for better mouse tooltip
|
|
// - goto declaration while in definition of recursive type
|
|
|
|
optional<QueryLocation> def_loc = GetDefinitionSpellingOfSymbol(db, ref.idx);
|
|
|
|
// We use spelling start and extent end because this causes vscode to
|
|
// highlight the entire definition when previewing / hoving with the
|
|
// mouse.
|
|
optional<QueryLocation> def_extent = GetDefinitionExtentOfSymbol(db, ref.idx);
|
|
if (def_loc && def_extent)
|
|
def_loc->range.end = def_extent->range.end;
|
|
|
|
// If the cursor is currently at or in the definition we should goto
|
|
// the declaration if possible. We also want to use declarations if
|
|
// we're pointing to, ie, a pure virtual function which has no
|
|
// definition.
|
|
if (!def_loc || (def_loc->path == file_id &&
|
|
def_loc->range.Contains(target_line, target_column))) {
|
|
// Goto declaration.
|
|
|
|
std::vector<QueryLocation> declarations = GetDeclarationsOfSymbolForGotoDefinition(db, ref.idx);
|
|
for (auto declaration : declarations) {
|
|
optional<lsLocation> ls_declaration = GetLsLocation(db, working_files, declaration);
|
|
if (ls_declaration)
|
|
response.result.push_back(*ls_declaration);
|
|
}
|
|
// We found some declarations. Break so we don't add the definition location.
|
|
if (!response.result.empty())
|
|
break;
|
|
}
|
|
|
|
if (def_loc)
|
|
PushBack(&response.result, GetLsLocation(db, working_files, *def_loc));
|
|
|
|
if (!response.result.empty())
|
|
break;
|
|
}
|
|
|
|
// No symbols - check for includes.
|
|
if (response.result.empty()) {
|
|
for (const IndexInclude& include : file->def.includes) {
|
|
if (include.line == target_line) {
|
|
lsLocation result;
|
|
result.uri = lsDocumentUri::FromPath(include.resolved_path);
|
|
response.result.push_back(result);
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
ipc->SendOutMessageToClient(IpcId::TextDocumentDefinition, response);
|
|
break;
|
|
}
|
|
|
|
case IpcId::TextDocumentDocumentHighlight: {
|
|
auto msg = static_cast<Ipc_TextDocumentDocumentHighlight*>(message.get());
|
|
|
|
QueryFileId file_id;
|
|
QueryFile* file = FindFile(db, msg->params.textDocument.uri.GetPath(), &file_id);
|
|
if (!file) {
|
|
std::cerr << "Unable to find file " << msg->params.textDocument.uri.GetPath() << std::endl;
|
|
break;
|
|
}
|
|
WorkingFile* working_file = working_files->GetFileByFilename(file->def.path);
|
|
|
|
Out_TextDocumentDocumentHighlight response;
|
|
response.id = msg->id;
|
|
|
|
for (const SymbolRef& ref : FindSymbolsAtLocation(working_file, file, msg->params.position)) {
|
|
// Found symbol. Return references to highlight.
|
|
std::vector<QueryLocation> uses = GetUsesOfSymbol(db, ref.idx);
|
|
response.result.reserve(uses.size());
|
|
for (const QueryLocation& use : uses) {
|
|
if (use.path != file_id)
|
|
continue;
|
|
|
|
optional<lsLocation> ls_location = GetLsLocation(db, working_files, use);
|
|
if (!ls_location)
|
|
continue;
|
|
|
|
lsDocumentHighlight highlight;
|
|
highlight.kind = lsDocumentHighlightKind::Text;
|
|
highlight.range = ls_location->range;
|
|
response.result.push_back(highlight);
|
|
}
|
|
break;
|
|
}
|
|
|
|
ipc->SendOutMessageToClient(IpcId::TextDocumentDocumentHighlight, response);
|
|
break;
|
|
}
|
|
|
|
case IpcId::TextDocumentHover: {
|
|
auto msg = static_cast<Ipc_TextDocumentHover*>(message.get());
|
|
|
|
QueryFile* file = FindFile(db, msg->params.textDocument.uri.GetPath());
|
|
if (!file) {
|
|
std::cerr << "Unable to find file " << msg->params.textDocument.uri.GetPath() << std::endl;
|
|
break;
|
|
}
|
|
WorkingFile* working_file = working_files->GetFileByFilename(file->def.path);
|
|
|
|
Out_TextDocumentHover response;
|
|
response.id = msg->id;
|
|
|
|
for (const SymbolRef& ref : FindSymbolsAtLocation(working_file, file, msg->params.position)) {
|
|
// Found symbol. Return hover.
|
|
optional<lsRange> ls_range = GetLsRange(working_files->GetFileByFilename(file->def.path), ref.loc.range);
|
|
if (!ls_range)
|
|
continue;
|
|
|
|
response.result.contents = GetHoverForSymbol(db, ref.idx);
|
|
response.result.range = *ls_range;
|
|
break;
|
|
}
|
|
|
|
ipc->SendOutMessageToClient(IpcId::TextDocumentHover, response);
|
|
break;
|
|
}
|
|
|
|
case IpcId::TextDocumentReferences: {
|
|
auto msg = static_cast<Ipc_TextDocumentReferences*>(message.get());
|
|
|
|
QueryFile* file = FindFile(db, msg->params.textDocument.uri.GetPath());
|
|
if (!file) {
|
|
std::cerr << "Unable to find file " << msg->params.textDocument.uri.GetPath() << std::endl;
|
|
break;
|
|
}
|
|
WorkingFile* working_file = working_files->GetFileByFilename(file->def.path);
|
|
|
|
Out_TextDocumentReferences response;
|
|
response.id = msg->id;
|
|
|
|
for (const SymbolRef& ref : FindSymbolsAtLocation(working_file, file, msg->params.position)) {
|
|
optional<QueryLocation> excluded_declaration;
|
|
if (!msg->params.context.includeDeclaration) {
|
|
std::cerr << "Excluding declaration in references" << std::endl;
|
|
excluded_declaration = GetDefinitionSpellingOfSymbol(db, ref.idx);
|
|
}
|
|
|
|
// Found symbol. Return references.
|
|
std::vector<QueryLocation> uses = GetUsesOfSymbol(db, ref.idx);
|
|
response.result.reserve(uses.size());
|
|
for (const QueryLocation& use : uses) {
|
|
if (excluded_declaration.has_value() && use == *excluded_declaration)
|
|
continue;
|
|
|
|
optional<lsLocation> ls_location = GetLsLocation(db, working_files, use);
|
|
if (ls_location)
|
|
response.result.push_back(*ls_location);
|
|
}
|
|
break;
|
|
}
|
|
|
|
ipc->SendOutMessageToClient(IpcId::TextDocumentReferences, response);
|
|
break;
|
|
}
|
|
|
|
case IpcId::TextDocumentDocumentSymbol: {
|
|
auto msg = static_cast<Ipc_TextDocumentDocumentSymbol*>(message.get());
|
|
|
|
Out_TextDocumentDocumentSymbol response;
|
|
response.id = msg->id;
|
|
|
|
QueryFile* file = FindFile(db, msg->params.textDocument.uri.GetPath());
|
|
if (!file) {
|
|
std::cerr << "Unable to find file " << msg->params.textDocument.uri.GetPath() << std::endl;
|
|
break;
|
|
}
|
|
|
|
for (SymbolRef ref : file->def.outline) {
|
|
optional<lsSymbolInformation> info = GetSymbolInfo(db, working_files, ref.idx);
|
|
if (!info)
|
|
continue;
|
|
|
|
optional<lsLocation> location = GetLsLocation(db, working_files, ref.loc);
|
|
if (!location)
|
|
continue;
|
|
info->location = *location;
|
|
response.result.push_back(*info);
|
|
}
|
|
|
|
ipc->SendOutMessageToClient(IpcId::TextDocumentDocumentSymbol, response);
|
|
break;
|
|
}
|
|
|
|
case IpcId::TextDocumentDocumentLink: {
|
|
auto msg = static_cast<Ipc_TextDocumentDocumentLink*>(message.get());
|
|
|
|
Out_TextDocumentDocumentLink response;
|
|
response.id = msg->id;
|
|
|
|
if (config->showDocumentLinksOnIncludes) {
|
|
QueryFile* file = FindFile(db, msg->params.textDocument.uri.GetPath());
|
|
if (!file) {
|
|
std::cerr << "Unable to find file " << msg->params.textDocument.uri.GetPath() << std::endl;
|
|
break;
|
|
}
|
|
|
|
WorkingFile* working_file = working_files->GetFileByFilename(msg->params.textDocument.uri.GetPath());
|
|
if (!working_file) {
|
|
std::cerr << "Unable to find working file " << msg->params.textDocument.uri.GetPath() << std::endl;
|
|
break;
|
|
}
|
|
for (const IndexInclude& include : file->def.includes) {
|
|
optional<int> buffer_line;
|
|
optional<std::string> buffer_line_content = working_file->GetBufferLineContentFromIndexLine(include.line, &buffer_line);
|
|
if (!buffer_line || !buffer_line_content)
|
|
continue;
|
|
|
|
// Find starting and ending quote.
|
|
int start = 0;
|
|
while (start < buffer_line_content->size()) {
|
|
char c = (*buffer_line_content)[start];
|
|
++start;
|
|
if (c == '"' || c == '<')
|
|
break;
|
|
}
|
|
if (start == buffer_line_content->size())
|
|
continue;
|
|
int end = (int)buffer_line_content->size();
|
|
while (end > 0) {
|
|
char c = (*buffer_line_content)[end];
|
|
if (c == '"' || c == '>')
|
|
break;
|
|
--end;
|
|
}
|
|
if (start >= end)
|
|
break;
|
|
|
|
|
|
lsDocumentLink link;
|
|
link.target = lsDocumentUri::FromPath(include.resolved_path);
|
|
// Subtract 1 from line because querydb stores 1-based lines but
|
|
// vscode expects 0-based lines.
|
|
link.range.start.line = *buffer_line - 1;
|
|
link.range.start.character = start;
|
|
link.range.end.line = *buffer_line - 1;
|
|
link.range.end.character = end;
|
|
response.result.push_back(link);
|
|
}
|
|
}
|
|
|
|
ipc->SendOutMessageToClient(IpcId::TextDocumentDocumentLink, response);
|
|
break;
|
|
}
|
|
|
|
case IpcId::TextDocumentCodeAction: {
|
|
// NOTE: This code snippet will generate some FixIts for testing:
|
|
//
|
|
// struct origin { int x, int y };
|
|
// void foo() {
|
|
// point origin = {
|
|
// x: 0.0,
|
|
// y: 0.0
|
|
// };
|
|
// }
|
|
//
|
|
auto msg = static_cast<Ipc_TextDocumentCodeAction*>(message.get());
|
|
|
|
WorkingFile* working_file = working_files->GetFileByFilename(msg->params.textDocument.uri.GetPath());
|
|
if (!working_file) {
|
|
// TODO: send error response.
|
|
std::cerr << "[error] textDocument/codeAction could not find working file" << std::endl;
|
|
break;
|
|
}
|
|
|
|
int target_line = msg->params.range.start.line;
|
|
|
|
Out_TextDocumentCodeAction response;
|
|
response.id = msg->id;
|
|
|
|
for (lsDiagnostic& diag : working_file->diagnostics) {
|
|
// clang does not provide accurate ennough column reporting for
|
|
// diagnostics to do good column filtering, so report all
|
|
// diagnostics on the line.
|
|
if (!diag.fixits_.empty() && diag.range.start.line == target_line) {
|
|
Out_TextDocumentCodeAction::Command command;
|
|
command.title = "FixIt: " + diag.message;
|
|
command.command = "cquery._applyFixIt";
|
|
command.arguments.textDocumentUri = msg->params.textDocument.uri;
|
|
command.arguments.edits = diag.fixits_;
|
|
response.result.push_back(command);
|
|
}
|
|
}
|
|
|
|
ipc->SendOutMessageToClient(IpcId::TextDocumentCodeAction, response);
|
|
break;
|
|
}
|
|
|
|
case IpcId::TextDocumentCodeLens: {
|
|
auto msg = static_cast<Ipc_TextDocumentCodeLens*>(message.get());
|
|
|
|
Out_TextDocumentCodeLens response;
|
|
response.id = msg->id;
|
|
|
|
lsDocumentUri file_as_uri = msg->params.textDocument.uri;
|
|
|
|
QueryFile* file = FindFile(db, file_as_uri.GetPath());
|
|
if (!file) {
|
|
std::cerr << "Unable to find file " << msg->params.textDocument.uri.GetPath() << std::endl;
|
|
break;
|
|
}
|
|
CommonCodeLensParams common;
|
|
common.result = &response.result;
|
|
common.db = db;
|
|
common.working_files = working_files;
|
|
common.working_file = working_files->GetFileByFilename(file->def.path);
|
|
|
|
for (SymbolRef ref : file->def.outline) {
|
|
// NOTE: We OffsetColumn so that the code lens always show up in a
|
|
// predictable order. Otherwise, the client may randomize it.
|
|
|
|
SymbolIdx symbol = ref.idx;
|
|
switch (symbol.kind) {
|
|
case SymbolKind::Type: {
|
|
optional<QueryType>& type = db->types[symbol.idx];
|
|
if (!type)
|
|
continue;
|
|
AddCodeLens("ref", "refs", &common, ref.loc.OffsetStartColumn(0), type->uses, type->def.definition_spelling, true /*force_display*/);
|
|
AddCodeLens("derived", "derived", &common, ref.loc.OffsetStartColumn(1), ToQueryLocation(db, type->derived), nullopt, false /*force_display*/);
|
|
AddCodeLens("var", "vars", &common, ref.loc.OffsetStartColumn(2), ToQueryLocation(db, type->instances), nullopt, false /*force_display*/);
|
|
break;
|
|
}
|
|
case SymbolKind::Func: {
|
|
optional<QueryFunc>& func = db->funcs[symbol.idx];
|
|
if (!func)
|
|
continue;
|
|
|
|
int16_t offset = 0;
|
|
|
|
std::vector<QueryFuncRef> base_callers = GetCallersForAllBaseFunctions(db, *func);
|
|
std::vector<QueryFuncRef> derived_callers = GetCallersForAllDerivedFunctions(db, *func);
|
|
if (base_callers.empty() && derived_callers.empty()) {
|
|
AddCodeLens("call", "calls", &common, ref.loc.OffsetStartColumn(offset++), ToQueryLocation(db, func->callers), nullopt, true /*force_display*/);
|
|
}
|
|
else {
|
|
AddCodeLens("direct call", "direct calls", &common, ref.loc.OffsetStartColumn(offset++), ToQueryLocation(db, func->callers), nullopt, false /*force_display*/);
|
|
if (!base_callers.empty())
|
|
AddCodeLens("base call", "base calls", &common, ref.loc.OffsetStartColumn(offset++), ToQueryLocation(db, base_callers), nullopt, false /*force_display*/);
|
|
if (!derived_callers.empty())
|
|
AddCodeLens("derived call", "derived calls", &common, ref.loc.OffsetStartColumn(offset++), ToQueryLocation(db, derived_callers), nullopt, false /*force_display*/);
|
|
}
|
|
|
|
AddCodeLens("derived", "derived", &common, ref.loc.OffsetStartColumn(offset++), ToQueryLocation(db, func->derived), nullopt, false /*force_display*/);
|
|
|
|
// "Base"
|
|
optional<QueryLocation> base_loc = GetBaseDefinitionOrDeclarationSpelling(db, *func);
|
|
if (base_loc) {
|
|
optional<lsLocation> ls_base = GetLsLocation(db, working_files, *base_loc);
|
|
if (ls_base) {
|
|
optional<lsRange> range = GetLsRange(common.working_file, ref.loc.range);
|
|
if (range) {
|
|
TCodeLens code_lens;
|
|
code_lens.range = *range;
|
|
code_lens.range.start.character += offset++;
|
|
code_lens.command = lsCommand<lsCodeLensCommandArguments>();
|
|
code_lens.command->title = "Base";
|
|
code_lens.command->command = "cquery.goto";
|
|
code_lens.command->arguments.uri = ls_base->uri;
|
|
code_lens.command->arguments.position = ls_base->range.start;
|
|
response.result.push_back(code_lens);
|
|
}
|
|
}
|
|
}
|
|
|
|
break;
|
|
}
|
|
case SymbolKind::Var: {
|
|
optional<QueryVar>& var = db->vars[symbol.idx];
|
|
if (!var)
|
|
continue;
|
|
|
|
if (var->def.is_local && !config->codeLensOnLocalVariables)
|
|
continue;
|
|
|
|
AddCodeLens("ref", "refs", &common, ref.loc.OffsetStartColumn(0), var->uses, var->def.definition_spelling, true /*force_display*/);
|
|
break;
|
|
}
|
|
case SymbolKind::File:
|
|
case SymbolKind::Invalid: {
|
|
assert(false && "unexpected");
|
|
break;
|
|
}
|
|
};
|
|
}
|
|
|
|
ipc->SendOutMessageToClient(IpcId::TextDocumentCodeLens, response);
|
|
break;
|
|
}
|
|
|
|
case IpcId::WorkspaceSymbol: {
|
|
// TODO: implement fuzzy search, see https://github.com/junegunn/fzf/blob/master/src/matcher.go for inspiration
|
|
auto msg = static_cast<Ipc_WorkspaceSymbol*>(message.get());
|
|
|
|
Out_WorkspaceSymbol response;
|
|
response.id = msg->id;
|
|
|
|
|
|
std::cerr << "[querydb] Considering " << db->detailed_names.size()
|
|
<< " candidates for query " << msg->params.query << std::endl;
|
|
|
|
std::string query = msg->params.query;
|
|
for (int i = 0; i < db->detailed_names.size(); ++i) {
|
|
if (response.result.size() >= config->maxWorkspaceSearchResults) {
|
|
std::cerr << "[querydb] Query exceeded maximum number of responses (" << config->maxWorkspaceSearchResults << "), output may not contain all results" << std::endl;
|
|
break;
|
|
}
|
|
|
|
if (db->detailed_names[i].find(query) != std::string::npos) {
|
|
optional<lsSymbolInformation> info = GetSymbolInfo(db, working_files, db->symbols[i]);
|
|
if (!info)
|
|
continue;
|
|
|
|
optional<QueryLocation> location = GetDefinitionExtentOfSymbol(db, db->symbols[i]);
|
|
if (!location) {
|
|
auto decls = GetDeclarationsOfSymbolForGotoDefinition(db, db->symbols[i]);
|
|
if (decls.empty())
|
|
continue;
|
|
location = decls[0];
|
|
}
|
|
|
|
optional<lsLocation> ls_location = GetLsLocation(db, working_files, *location);
|
|
if (!ls_location)
|
|
continue;
|
|
info->location = *ls_location;
|
|
response.result.push_back(*info);
|
|
}
|
|
}
|
|
|
|
std::cerr << "[querydb] Found " << response.result.size() << " results for query " << query << std::endl;
|
|
ipc->SendOutMessageToClient(IpcId::WorkspaceSymbol, response);
|
|
break;
|
|
}
|
|
|
|
default: {
|
|
std::cerr << "[querydb] Unhandled IPC message " << IpcIdToString(message->method_id) << std::endl;
|
|
exit(1);
|
|
}
|
|
}
|
|
}
|
|
|
|
// TODO: consider rate-limiting and checking for IPC messages so we don't block
|
|
// requests / we can serve partial requests.
|
|
|
|
|
|
while (true) {
|
|
optional<Index_DoIdMap> request = queue_do_id_map->TryDequeue();
|
|
if (!request)
|
|
break;
|
|
|
|
did_work = true;
|
|
|
|
Index_OnIdMapped response(request->indexed_content, request->perf, request->is_interactive);
|
|
Timer time;
|
|
|
|
if (request->previous) {
|
|
response.previous_id_map = MakeUnique<IdMap>(db, request->previous->id_cache);
|
|
response.previous_index = std::move(request->previous);
|
|
}
|
|
|
|
assert(request->current);
|
|
response.current_id_map = MakeUnique<IdMap>(db, request->current->id_cache);
|
|
response.current_index = std::move(request->current);
|
|
response.perf.querydb_id_map = time.ElapsedMicrosecondsAndReset();
|
|
|
|
queue_on_id_mapped->Enqueue(std::move(response));
|
|
}
|
|
|
|
while (true) {
|
|
optional<Index_OnIndexed> response = queue_on_indexed->TryDequeue();
|
|
if (!response)
|
|
break;
|
|
|
|
did_work = true;
|
|
|
|
Timer time;
|
|
|
|
for (auto& updated_file : response->update.files_def_update) {
|
|
// TODO: We're reading a file on querydb thread. This is slow!! If it is a
|
|
// problem in practice we need to create a file reader queue, dispatch the
|
|
// read to it, get a response, and apply the new index then.
|
|
WorkingFile* working_file = working_files->GetFileByFilename(updated_file.path);
|
|
if (working_file) {
|
|
auto it = response->indexed_content.find(updated_file.path);
|
|
if (it != response->indexed_content.end()) {
|
|
working_file->SetIndexContent(it->second);
|
|
time.ResetAndPrint("[querydb] Update WorkingFile index contents (via in-memory buffer) for " + updated_file.path);
|
|
}
|
|
else {
|
|
optional<std::string> cached_file_contents = LoadCachedFileContents(config, updated_file.path);
|
|
if (cached_file_contents)
|
|
working_file->SetIndexContent(*cached_file_contents);
|
|
else
|
|
working_file->SetIndexContent(working_file->buffer_content);
|
|
time.ResetAndPrint("[querydb] Update WorkingFile index contents (via disk load) for " + updated_file.path);
|
|
}
|
|
}
|
|
}
|
|
|
|
db->ApplyIndexUpdate(&response->update);
|
|
//time.ResetAndPrint("[querydb] Applying index update");
|
|
}
|
|
|
|
return did_work;
|
|
}
|
|
|
|
void QueryDbMain(Config* config, MultiQueueWaiter* waiter) {
|
|
// Create queues.
|
|
Index_DoIndexQueue queue_do_index(waiter);
|
|
Index_DoIdMapQueue queue_do_id_map(waiter);
|
|
Index_OnIdMappedQueue queue_on_id_mapped(waiter);
|
|
Index_OnIndexedQueue queue_on_indexed(waiter);
|
|
|
|
Project project;
|
|
WorkingFiles working_files;
|
|
CompletionManager completion_manager(config, &project, &working_files);
|
|
IncludeCompletion include_completion(config, &project);
|
|
CodeCompleteCache code_complete_cache;
|
|
CodeCompleteCache signature_cache;
|
|
FileConsumer::SharedState file_consumer_shared;
|
|
|
|
// Run query db main loop.
|
|
SetCurrentThreadName("querydb");
|
|
QueryDatabase db;
|
|
while (true) {
|
|
bool did_work = QueryDbMainLoop(
|
|
config, &db, waiter, &queue_do_index, &queue_do_id_map, &queue_on_id_mapped, &queue_on_indexed,
|
|
&project, &file_consumer_shared, &working_files,
|
|
&completion_manager, &include_completion, &code_complete_cache, &signature_cache);
|
|
if (!did_work) {
|
|
IpcManager* ipc = IpcManager::instance();
|
|
waiter->Wait({
|
|
ipc->threaded_queue_for_server_.get(),
|
|
&queue_do_id_map,
|
|
&queue_on_indexed
|
|
});
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
// TODO: global lock on stderr output.
|
|
|
|
// Separate thread whose only job is to read from stdin and
|
|
// dispatch read commands to the actual indexer program. This
|
|
// cannot be done on the main thread because reading from std::cin
|
|
// blocks.
|
|
//
|
|
// |ipc| is connected to a server.
|
|
void LanguageServerStdinLoop(Config* config, std::unordered_map<IpcId, Timer>* request_times) {
|
|
IpcManager* ipc = IpcManager::instance();
|
|
|
|
SetCurrentThreadName("stdin");
|
|
while (true) {
|
|
std::unique_ptr<BaseIpcMessage> message = MessageRegistry::instance()->ReadMessageFromStdin();
|
|
|
|
// Message parsing can fail if we don't recognize the method.
|
|
if (!message)
|
|
continue;
|
|
|
|
(*request_times)[message->method_id] = Timer();
|
|
|
|
//std::cerr << "[stdin] Got message " << IpcIdToString(message->method_id) << std::endl;
|
|
switch (message->method_id) {
|
|
case IpcId::Initialized: {
|
|
// TODO: don't send output until we get this notification
|
|
break;
|
|
}
|
|
|
|
case IpcId::CancelRequest: {
|
|
// TODO: support cancellation
|
|
break;
|
|
}
|
|
|
|
case IpcId::Initialize:
|
|
case IpcId::Exit:
|
|
case IpcId::TextDocumentDidOpen:
|
|
case IpcId::TextDocumentDidChange:
|
|
case IpcId::TextDocumentDidClose:
|
|
case IpcId::TextDocumentDidSave:
|
|
case IpcId::TextDocumentRename:
|
|
case IpcId::TextDocumentCompletion:
|
|
case IpcId::TextDocumentSignatureHelp:
|
|
case IpcId::TextDocumentDefinition:
|
|
case IpcId::TextDocumentDocumentHighlight:
|
|
case IpcId::TextDocumentHover:
|
|
case IpcId::TextDocumentReferences:
|
|
case IpcId::TextDocumentDocumentSymbol:
|
|
case IpcId::TextDocumentDocumentLink:
|
|
case IpcId::TextDocumentCodeAction:
|
|
case IpcId::TextDocumentCodeLens:
|
|
case IpcId::WorkspaceSymbol:
|
|
case IpcId::CqueryFreshenIndex:
|
|
case IpcId::CqueryVars:
|
|
case IpcId::CqueryCallers:
|
|
case IpcId::CqueryBase:
|
|
case IpcId::CqueryDerived: {
|
|
ipc->SendMessage(IpcManager::Destination::Server, std::move(message));
|
|
break;
|
|
}
|
|
|
|
default: {
|
|
std::cerr << "[stdin] Unhandled IPC message " << IpcIdToString(message->method_id) << std::endl;
|
|
exit(1);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
void StdoutMain(std::unordered_map<IpcId, Timer>* request_times, MultiQueueWaiter* waiter) {
|
|
SetCurrentThreadName("stdout");
|
|
IpcManager* ipc = IpcManager::instance();
|
|
|
|
while (true) {
|
|
std::vector<std::unique_ptr<BaseIpcMessage>> messages = ipc->GetMessages(IpcManager::Destination::Client);
|
|
if (messages.empty()) {
|
|
waiter->Wait({ipc->threaded_queue_for_client_.get()});
|
|
continue;
|
|
}
|
|
|
|
for (auto& message : messages) {
|
|
//std::cerr << "[stdout] Processing message " << IpcIdToString(message->method_id) << std::endl;
|
|
|
|
switch (message->method_id) {
|
|
case IpcId::Cout: {
|
|
auto msg = static_cast<Ipc_Cout*>(message.get());
|
|
|
|
if (ShouldDisplayIpcTiming(message->method_id)) {
|
|
Timer time = (*request_times)[msg->original_ipc_id];
|
|
time.ResetAndPrint("[e2e] Running " + std::string(IpcIdToString(msg->original_ipc_id)));
|
|
}
|
|
|
|
std::cout << msg->content;
|
|
std::cout.flush();
|
|
break;
|
|
}
|
|
|
|
default: {
|
|
std::cerr << "[stdout] Unhandled IPC message " << IpcIdToString(message->method_id) << std::endl;
|
|
exit(1);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
void LanguageServerMain(Config* config, MultiQueueWaiter* waiter) {
|
|
std::unordered_map<IpcId, Timer> request_times;
|
|
|
|
// Start stdin reader. Reading from stdin is a blocking operation so this
|
|
// needs a dedicated thread.
|
|
new std::thread([&]() {
|
|
LanguageServerStdinLoop(config, &request_times);
|
|
});
|
|
|
|
// Start querydb thread. querydb will start indexer threads as needed.
|
|
new std::thread([&]() {
|
|
QueryDbMain(config, waiter);
|
|
});
|
|
|
|
// We run a dedicated thread for writing to stdout because there can be an
|
|
// unknown number of delays when output information.
|
|
StdoutMain(&request_times, waiter);
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
int main(int argc, char** argv) {
|
|
MultiQueueWaiter waiter;
|
|
IpcManager::CreateInstance(&waiter);
|
|
|
|
//bool loop = true;
|
|
//while (loop)
|
|
// std::this_thread::sleep_for(std::chrono::milliseconds(10));
|
|
//std::this_thread::sleep_for(std::chrono::seconds(3));
|
|
|
|
PlatformInit();
|
|
IndexInit();
|
|
|
|
RegisterMessageTypes();
|
|
|
|
std::unordered_map<std::string, std::string> options =
|
|
ParseOptions(argc, argv);
|
|
|
|
if (HasOption(options, "--test")) {
|
|
doctest::Context context;
|
|
context.applyCommandLine(argc, argv);
|
|
int res = context.run();
|
|
if (context.shouldExit())
|
|
return res;
|
|
|
|
RunTests();
|
|
return 0;
|
|
}
|
|
else if (HasOption(options, "--language-server")) {
|
|
//std::cerr << "Running language server" << std::endl;
|
|
Config config;
|
|
LanguageServerMain(&config, &waiter);
|
|
return 0;
|
|
}
|
|
else {
|
|
std::cout << R"help(cquery help:
|
|
|
|
cquery is a low-latency C++ language server.
|
|
|
|
General:
|
|
--help Print this help information.
|
|
--language-server
|
|
Run as a language server. This implements the language
|
|
server spec over STDIN and STDOUT.
|
|
--test Run tests. Does nothing if test support is not compiled in.
|
|
|
|
Configuration:
|
|
When opening up a directory, cquery will look for a compile_commands.json
|
|
file emitted by your preferred build system. If not present, cquery will
|
|
use a recursive directory listing instead. Command line flags can be
|
|
provided by adding a "clang_args" file in the top-level directory. Each
|
|
line in that file is a separate argument.
|
|
|
|
There are also a number of configuration options available when
|
|
initializing the language server - your editor should have tooling to
|
|
describe those options. See |Config| in this source code for a detailed
|
|
list of all currently supported options.
|
|
)help";
|
|
return 0;
|
|
}
|
|
}
|