// TODO: cleanup includes #include "cache.h" #include "cache_loader.h" #include "clang_complete.h" #include "file_consumer.h" #include "include_complete.h" #include "indexer.h" #include "ipc_manager.h" #include "language_server_api.h" #include "lex_utils.h" #include "lru_cache.h" #include "match.h" #include "message_handler.h" #include "options.h" #include "platform.h" #include "project.h" #include "query.h" #include "query_utils.h" #include "serializer.h" #include "standard_includes.h" #include "test.h" #include "threaded_queue.h" #include "timer.h" #include "timestamp_manager.h" #include "work_thread.h" #include "working_files.h" #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include // TODO: provide a feature like 'https://github.com/goldsborough/clang-expand', // ie, a fully linear view of a function with inline function calls expanded. // We can probably use vscode decorators to achieve it. // TODO: implement ThreadPool type which monitors CPU usage / number of work // items per second completed and scales up/down number of running threads. namespace { std::vector kEmptyArgs; // If true stdout will be printed to stderr. bool g_log_stdin_stdout_to_stderr = false; // This function returns true if e2e timing should be displayed for the given // IpcId. bool ShouldDisplayIpcTiming(IpcId id) { switch (id) { case IpcId::TextDocumentPublishDiagnostics: case IpcId::CqueryPublishInactiveRegions: case IpcId::Unknown: return false; default: return true; } } void EmitDiagnostics(WorkingFiles* working_files, std::string path, std::vector diagnostics) { // Emit diagnostics. Out_TextDocumentPublishDiagnostics out; out.params.uri = lsDocumentUri::FromPath(path); out.params.diagnostics = diagnostics; IpcManager::WriteStdout(IpcId::TextDocumentPublishDiagnostics, out); // Cache diagnostics so we can show fixits. working_files->DoActionOnFile(path, [&](WorkingFile* working_file) { if (working_file) working_file->diagnostics_ = diagnostics; }); } REGISTER_IPC_MESSAGE(Ipc_CancelRequest); // Send indexing progress to client if reporting is enabled. void EmitProgress(Config* config, QueueManager* queue) { if (config->enableProgressReports) { Out_Progress out; out.params.indexRequestCount = queue->index_request.Size(); out.params.doIdMapCount = queue->do_id_map.Size(); out.params.loadPreviousIndexCount = queue->load_previous_index.Size(); out.params.onIdMappedCount = queue->on_id_mapped.Size(); out.params.onIndexedCount = queue->on_indexed.Size(); IpcManager::WriteStdout(IpcId::Unknown, out); } } } // namespace //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// // IMPORT PIPELINE ///////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// enum class FileParseQuery { NeedsParse, DoesNotNeedParse, NoSuchFile }; std::vector DoParseFile( Config* config, WorkingFiles* working_files, ClangIndex* index, FileConsumer::SharedState* file_consumer_shared, TimestampManager* timestamp_manager, ImportManager* import_manager, CacheLoader* cache_loader, bool is_interactive, const std::string& path, const std::vector& args, const optional& contents) { std::vector result; // Always run this block, even if we are interactive, so we can check // dependencies and reset files in |file_consumer_shared|. IndexFile* previous_index = cache_loader->TryLoad(path); if (previous_index) { // If none of the dependencies have changed and the index is not // interactive (ie, requested by a file save), skip parsing and just load // from cache. // Checks if |path| needs to be reparsed. This will modify cached state // such that calling this function twice with the same path may return true // the first time but will return false the second. auto file_needs_parse = [&](const std::string& path, bool is_dependency) { // If the file is a dependency but another file as already imported it, // don't bother. if (!is_interactive && is_dependency && !import_manager->TryMarkDependencyImported(path)) { return FileParseQuery::DoesNotNeedParse; } optional modification_timestamp = GetLastModificationTime(path); // Cannot find file. if (!modification_timestamp) return FileParseQuery::NoSuchFile; optional last_cached_modification = timestamp_manager->GetLastCachedModificationTime(cache_loader, path); // File has been changed. if (!last_cached_modification || modification_timestamp != *last_cached_modification) { file_consumer_shared->Reset(path); timestamp_manager->UpdateCachedModificationTime( path, *modification_timestamp); return FileParseQuery::NeedsParse; } // File has not changed, do not parse it. return FileParseQuery::DoesNotNeedParse; }; // Check timestamps and update |file_consumer_shared|. FileParseQuery path_state = file_needs_parse(path, false /*is_dependency*/); // Target file does not exist on disk, do not emit any indexes. // TODO: Dependencies should be reassigned to other files. We can do this by // updating the "primary_file" if it doesn't exist. Might not actually be a // problem in practice. if (path_state == FileParseQuery::NoSuchFile) return result; bool needs_reparse = is_interactive || path_state == FileParseQuery::NeedsParse; for (const std::string& dependency : previous_index->dependencies) { assert(!dependency.empty()); // note: Use != as there are multiple failure results for FileParseQuery. if (file_needs_parse(dependency, true /*is_dependency*/) != FileParseQuery::DoesNotNeedParse) { LOG_S(INFO) << "Timestamp has changed for " << dependency << " (via " << previous_index->path << ")"; needs_reparse = true; // SUBTLE: Do not break here, as |file_consumer_shared| is updated // inside of |file_needs_parse|. } } // No timestamps changed - load directly from cache. if (!needs_reparse) { LOG_S(INFO) << "Skipping parse; no timestamp change for " << path; // TODO/FIXME: real perf PerformanceImportFile perf; result.push_back(Index_DoIdMap(cache_loader->TakeOrLoad(path), perf, is_interactive, false /*write_to_disk*/)); for (const std::string& dependency : previous_index->dependencies) { // Only load a dependency if it is not already loaded. // // This is important for perf in large projects where there are lots of // dependencies shared between many files. if (!file_consumer_shared->Mark(dependency)) continue; LOG_S(INFO) << "Emitting index result for " << dependency << " (via " << previous_index->path << ")"; std::unique_ptr dependency_index = cache_loader->TryTakeOrLoad(dependency); // |dependency_index| may be null if there is no cache for it but // another file has already started importing it. if (!dependency_index) continue; result.push_back(Index_DoIdMap(std::move(dependency_index), perf, is_interactive, false /*write_to_disk*/)); } return result; } } LOG_S(INFO) << "Parsing " << path; // Load file contents for all dependencies into memory. If the dependencies // for the file changed we may not end up using all of the files we // preloaded. If a new dependency was added the indexer will grab the file // contents as soon as possible. // // We do this to minimize the race between indexing a file and capturing the // file contents. // // TODO: We might be able to optimize perf by only copying for files in // working_files. We can pass that same set of files to the indexer as // well. We then default to a fast file-copy if not in working set. bool loaded_primary = false; std::vector file_contents; if (contents) { loaded_primary = loaded_primary || contents->path == path; file_contents.push_back(*contents); } for (const auto& it : cache_loader->caches) { const std::unique_ptr& index = it.second; assert(index); optional index_content = ReadContent(index->path); if (!index_content) { LOG_S(ERROR) << "Failed to preload index content for " << index->path; continue; } file_contents.push_back(FileContents(index->path, *index_content)); loaded_primary = loaded_primary || index->path == path; } if (!loaded_primary) { optional content = ReadContent(path); if (!content) { LOG_S(ERROR) << "Skipping index (file cannot be found): " << path; return result; } file_contents.push_back(FileContents(path, *content)); } PerformanceImportFile perf; std::vector> indexes = Parse( config, file_consumer_shared, path, args, file_contents, &perf, index); for (std::unique_ptr& new_index : indexes) { Timer time; // Only emit diagnostics for non-interactive sessions, which makes it easier // to identify indexing problems. For interactive sessions, diagnostics are // handled by code completion. if (!is_interactive) EmitDiagnostics(working_files, new_index->path, new_index->diagnostics_); // When main thread does IdMap request it will request the previous index if // needed. LOG_S(INFO) << "Emitting index result for " << new_index->path; result.push_back(Index_DoIdMap(std::move(new_index), perf, is_interactive, true /*write_to_disk*/)); } return result; } // Index a file using an already-parsed translation unit from code completion. // Since most of the time for indexing a file comes from parsing, we can do // real-time indexing. // TODO: add option to disable this. void IndexWithTuFromCodeCompletion( QueueManager* queue, FileConsumer::SharedState* file_consumer_shared, ClangTranslationUnit* tu, const std::vector& file_contents, const std::string& path, const std::vector& args) { file_consumer_shared->Reset(path); PerformanceImportFile perf; ClangIndex index; std::vector> indexes = ParseWithTu( file_consumer_shared, &perf, tu, &index, path, args, file_contents); std::vector result; for (std::unique_ptr& new_index : indexes) { Timer time; // When main thread does IdMap request it will request the previous index if // needed. LOG_S(INFO) << "Emitting index result for " << new_index->path; result.push_back(Index_DoIdMap(std::move(new_index), perf, true /*is_interactive*/, true /*write_to_disk*/)); } LOG_IF_S(WARNING, result.size() > 1) << "Code completion index update generated more than one index"; queue->do_id_map.EnqueueAll(std::move(result)); } std::vector ParseFile( Config* config, WorkingFiles* working_files, ClangIndex* index, FileConsumer::SharedState* file_consumer_shared, TimestampManager* timestamp_manager, ImportManager* import_manager, bool is_interactive, const Project::Entry& entry, const optional& contents) { optional file_contents; if (contents) file_contents = FileContents(entry.filename, *contents); CacheLoader cache_loader(config); // Try to determine the original import file by loading the file from cache. // This lets the user request an index on a header file, which clang will // complain about if indexed by itself. IndexFile* entry_cache = cache_loader.TryLoad(entry.filename); std::string tu_path = entry_cache ? entry_cache->import_file : entry.filename; return DoParseFile(config, working_files, index, file_consumer_shared, timestamp_manager, import_manager, &cache_loader, is_interactive, tu_path, entry.args, file_contents); } bool IndexMain_DoParse(Config* config, WorkingFiles* working_files, QueueManager* queue, FileConsumer::SharedState* file_consumer_shared, TimestampManager* timestamp_manager, ImportManager* import_manager, ClangIndex* index) { optional request = queue->index_request.TryDequeue(); if (!request) return false; Project::Entry entry; entry.filename = request->path; entry.args = request->args; std::vector responses = ParseFile( config, working_files, index, file_consumer_shared, timestamp_manager, import_manager, request->is_interactive, entry, request->contents); // Don't bother sending an IdMap request if there are no responses. if (responses.empty()) return false; // EnqueueAll will clear |responses|. queue->do_id_map.EnqueueAll(std::move(responses)); return true; } bool IndexMain_DoCreateIndexUpdate(Config* config, QueueManager* queue, TimestampManager* timestamp_manager) { optional response = queue->on_id_mapped.TryDequeue(); if (!response) return false; Timer time; IdMap* previous_id_map = nullptr; IndexFile* previous_index = nullptr; if (response->previous) { previous_id_map = response->previous->ids.get(); previous_index = response->previous->file.get(); } // Build delta update. IndexUpdate update = IndexUpdate::CreateDelta(previous_id_map, response->current->ids.get(), previous_index, response->current->file.get()); response->perf.index_make_delta = time.ElapsedMicrosecondsAndReset(); LOG_S(INFO) << "Built index update for " << response->current->file->path << " (is_delta=" << !!response->previous << ")"; // Write current index to disk if requested. if (response->write_to_disk) { LOG_S(INFO) << "Writing cached index to disk for " << response->current->file->path; time.Reset(); WriteToCache(config, *response->current->file); response->perf.index_save_to_disk = time.ElapsedMicrosecondsAndReset(); timestamp_manager->UpdateCachedModificationTime( response->current->file->path, response->current->file->last_modification_time); } #if false #define PRINT_SECTION(name) \ if (response->perf.name) { \ total += response->perf.name; \ output << " " << #name << ": " << FormatMicroseconds(response->perf.name); \ } std::stringstream output; long long total = 0; output << "[perf]"; PRINT_SECTION(index_parse); PRINT_SECTION(index_build); PRINT_SECTION(index_save_to_disk); PRINT_SECTION(index_load_cached); PRINT_SECTION(querydb_id_map); PRINT_SECTION(index_make_delta); output << "\n total: " << FormatMicroseconds(total); output << " path: " << response->current_index->path; LOG_S(INFO) << output.rdbuf(); #undef PRINT_SECTION if (response->is_interactive) LOG_S(INFO) << "Applying IndexUpdate" << std::endl << update.ToString(); #endif Index_OnIndexed reply(update, response->perf); queue->on_indexed.Enqueue(std::move(reply)); return true; } bool IndexMain_LoadPreviousIndex(Config* config, QueueManager* queue) { optional response = queue->load_previous_index.TryDequeue(); if (!response) return false; response->previous = LoadCachedIndex(config, response->current->path); LOG_IF_S(ERROR, !response->previous) << "Unable to load previous index for already imported index " << response->current->path; queue->do_id_map.Enqueue(std::move(*response)); return true; } bool IndexMergeIndexUpdates(QueueManager* queue) { optional root = queue->on_indexed.TryDequeue(); if (!root) return false; bool did_merge = false; while (true) { optional to_join = queue->on_indexed.TryDequeue(); if (!to_join) { queue->on_indexed.Enqueue(std::move(*root)); return did_merge; } did_merge = true; Timer time; root->update.Merge(to_join->update); // time.ResetAndPrint("Joined querydb updates for files: " + // StringJoinMap(root->update.files_def_update, //[](const QueryFile::DefUpdate& update) { // return update.path; //})); } } WorkThread::Result IndexMain(Config* config, FileConsumer::SharedState* file_consumer_shared, TimestampManager* timestamp_manager, ImportManager* import_manager, Project* project, WorkingFiles* working_files, MultiQueueWaiter* waiter, QueueManager* queue) { EmitProgress(config, queue); // TODO: dispose of index after it is not used for a while. ClangIndex index; // TODO: process all off IndexMain_DoIndex before calling // IndexMain_DoCreateIndexUpdate for // better icache behavior. We need to have some threads spinning on // both though // otherwise memory usage will get bad. // We need to make sure to run both IndexMain_DoParse and // IndexMain_DoCreateIndexUpdate so we don't starve querydb from doing any // work. Running both also lets the user query the partially constructed // index. bool did_parse = IndexMain_DoParse(config, working_files, queue, file_consumer_shared, timestamp_manager, import_manager, &index); bool did_create_update = IndexMain_DoCreateIndexUpdate(config, queue, timestamp_manager); bool did_load_previous = IndexMain_LoadPreviousIndex(config, queue); // Nothing to index and no index updates to create, so join some already // created index updates to reduce work on querydb thread. bool did_merge = false; if (!did_parse && !did_create_update && !did_load_previous) did_merge = IndexMergeIndexUpdates(queue); // We didn't do any work, so wait for a notification. if (!did_parse && !did_create_update && !did_merge && !did_load_previous) { waiter->Wait({&queue->index_request, &queue->on_id_mapped, &queue->load_previous_index, &queue->on_indexed}); } return queue->HasWork() ? WorkThread::Result::MoreWork : WorkThread::Result::NoWork; } bool QueryDb_ImportMain(Config* config, QueryDatabase* db, ImportManager* import_manager, QueueManager* queue, SemanticHighlightSymbolCache* semantic_cache, WorkingFiles* working_files) { EmitProgress(config, queue); bool did_work = false; while (true) { optional request = queue->do_id_map.TryDequeue(); if (!request) break; did_work = true; assert(request->current); // If the request does not have previous state and we have already imported // it, load the previous state from disk and rerun IdMap logic later. Do not // do this if we have already attempted in the past. if (!request->load_previous && !request->previous && db->usr_to_file.find(LowerPathIfCaseInsensitive( request->current->path)) != db->usr_to_file.end()) { assert(!request->load_previous); request->load_previous = true; queue->load_previous_index.Enqueue(std::move(*request)); continue; } // Check if the file is already being imported into querydb. If it is, drop // the request. // // Note, we must do this *after* we have checked for the previous index, // otherwise we will never actually generate the IdMap. if (!import_manager->StartQueryDbImport(request->current->path)) { LOG_S(INFO) << "Dropping index as it is already being imported for " << request->current->path; continue; } Index_OnIdMapped response(request->perf, request->is_interactive, request->write_to_disk); Timer time; auto make_map = [db](std::unique_ptr file) -> std::unique_ptr { if (!file) return nullptr; auto id_map = MakeUnique(db, file->id_cache); return MakeUnique(std::move(file), std::move(id_map)); }; response.current = make_map(std::move(request->current)); response.previous = make_map(std::move(request->previous)); response.perf.querydb_id_map = time.ElapsedMicrosecondsAndReset(); queue->on_id_mapped.Enqueue(std::move(response)); } while (true) { optional response = queue->on_indexed.TryDequeue(); if (!response) break; did_work = true; Timer time; for (auto& updated_file : response->update.files_def_update) { // TODO: We're reading a file on querydb thread. This is slow!! If this // a real problem in practice we can load the file in a previous stage. // It should be fine though because we only do it if the user has the // file open. WorkingFile* working_file = working_files->GetFileByFilename(updated_file.path); if (working_file) { optional cached_file_contents = LoadCachedFileContents(config, updated_file.path); if (cached_file_contents) working_file->SetIndexContent(*cached_file_contents); else working_file->SetIndexContent(working_file->buffer_content); time.ResetAndPrint( "Update WorkingFile index contents (via disk load) for " + updated_file.path); // Update inactive region. EmitInactiveLines(working_file, updated_file.inactive_regions); } } time.Reset(); db->ApplyIndexUpdate(&response->update); time.ResetAndPrint("Applying index update for " + StringJoinMap(response->update.files_def_update, [](const QueryFile::DefUpdate& value) { return value.path; })); // Update semantic highlighting. for (auto& updated_file : response->update.files_def_update) { WorkingFile* working_file = working_files->GetFileByFilename(updated_file.path); if (working_file) { QueryFileId file_id = db->usr_to_file[LowerPathIfCaseInsensitive(working_file->filename)]; QueryFile* file = &db->files[file_id.id]; EmitSemanticHighlighting(db, semantic_cache, working_file, file); } } // Mark the files as being done in querydb stage after we apply the index // update. for (auto& updated_file : response->update.files_def_update) import_manager->DoneQueryDbImport(updated_file.path); } return did_work; } //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// // QUERYDB MAIN //////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// bool QueryDbMainLoop(Config* config, QueryDatabase* db, bool* exit_when_idle, MultiQueueWaiter* waiter, QueueManager* queue, Project* project, FileConsumer::SharedState* file_consumer_shared, ImportManager* import_manager, TimestampManager* timestamp_manager, SemanticHighlightSymbolCache* semantic_cache, WorkingFiles* working_files, ClangCompleteManager* clang_complete, IncludeComplete* include_complete, CodeCompleteCache* global_code_complete_cache, CodeCompleteCache* non_global_code_complete_cache, CodeCompleteCache* signature_cache) { IpcManager* ipc = IpcManager::instance(); bool did_work = false; std::vector> messages = ipc->for_querydb.DequeueAll(); for (auto& message : messages) { did_work = true; for (MessageHandler* handler : *MessageHandler::message_handlers) { if (handler->GetId() == message->method_id) { handler->Run(std::move(message)); break; } } if (message) { LOG_S(FATAL) << "Exiting; unhandled IPC message " << IpcIdToString(message->method_id); exit(1); } } // TODO: consider rate-limiting and checking for IPC messages so we don't // block requests / we can serve partial requests. if (QueryDb_ImportMain(config, db, import_manager, queue, semantic_cache, working_files)) { did_work = true; } return did_work; } void RunQueryDbThread(const std::string& bin_name, Config* config, MultiQueueWaiter* waiter, QueueManager* queue) { bool exit_when_idle = false; Project project; SemanticHighlightSymbolCache semantic_cache; WorkingFiles working_files; FileConsumer::SharedState file_consumer_shared; ClangCompleteManager clang_complete( config, &project, &working_files, std::bind(&EmitDiagnostics, &working_files, std::placeholders::_1, std::placeholders::_2), std::bind(&IndexWithTuFromCodeCompletion, queue, &file_consumer_shared, std::placeholders::_1, std::placeholders::_2, std::placeholders::_3, std::placeholders::_4)); IncludeComplete include_complete(config, &project); auto global_code_complete_cache = MakeUnique(); auto non_global_code_complete_cache = MakeUnique(); auto signature_cache = MakeUnique(); ImportManager import_manager; TimestampManager timestamp_manager; QueryDatabase db; // Setup shared references. for (MessageHandler* handler : *MessageHandler::message_handlers) { handler->config = config; handler->db = &db; handler->exit_when_idle = &exit_when_idle; handler->waiter = waiter; handler->queue = queue; handler->project = &project; handler->file_consumer_shared = &file_consumer_shared; handler->import_manager = &import_manager; handler->timestamp_manager = ×tamp_manager; handler->semantic_cache = &semantic_cache; handler->working_files = &working_files; handler->clang_complete = &clang_complete; handler->include_complete = &include_complete; handler->global_code_complete_cache = global_code_complete_cache.get(); handler->non_global_code_complete_cache = non_global_code_complete_cache.get(); handler->signature_cache = signature_cache.get(); } // Run query db main loop. SetCurrentThreadName("querydb"); while (true) { bool did_work = QueryDbMainLoop( config, &db, &exit_when_idle, waiter, queue, &project, &file_consumer_shared, &import_manager, ×tamp_manager, &semantic_cache, &working_files, &clang_complete, &include_complete, global_code_complete_cache.get(), non_global_code_complete_cache.get(), signature_cache.get()); // No more work left and exit request. Exit. if (!did_work && exit_when_idle && WorkThread::num_active_threads == 0) { LOG_S(INFO) << "Exiting; exit_when_idle is set and there is no more work"; exit(0); } // Cleanup and free any unused memory. FreeUnusedMemory(); if (!did_work) { waiter->Wait({&IpcManager::instance()->for_querydb, &queue->do_id_map, &queue->on_indexed}); } } } //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// // STDIN MAIN ////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// // Separate thread whose only job is to read from stdin and // dispatch read commands to the actual indexer program. This // cannot be done on the main thread because reading from std::cin // blocks. // // |ipc| is connected to a server. void LaunchStdinLoop(Config* config, std::unordered_map* request_times) { // If flushing cin requires flushing cout there could be deadlocks in some // clients. std::cin.tie(nullptr); WorkThread::StartThread("stdin", [request_times]() { IpcManager* ipc = IpcManager::instance(); std::unique_ptr message = MessageRegistry::instance()->ReadMessageFromStdin( g_log_stdin_stdout_to_stderr); // Message parsing can fail if we don't recognize the method. if (!message) return WorkThread::Result::MoreWork; (*request_times)[message->method_id] = Timer(); switch (message->method_id) { case IpcId::Initialized: { // TODO: don't send output until we get this notification break; } case IpcId::CancelRequest: { // TODO: support cancellation break; } case IpcId::Exit: { LOG_S(INFO) << "Exiting"; exit(0); break; } case IpcId::CqueryExitWhenIdle: { // querydb needs to know to exit when idle. We return out of the stdin // loop to exit the thread. If we keep parsing input stdin is likely // closed so cquery will exit. LOG_S(INFO) << "cquery will exit when all threads are idle"; ipc->for_querydb.Enqueue(std::move(message)); return WorkThread::Result::ExitThread; } case IpcId::Initialize: case IpcId::TextDocumentDidOpen: case IpcId::CqueryTextDocumentDidView: case IpcId::TextDocumentDidChange: case IpcId::TextDocumentDidClose: case IpcId::TextDocumentDidSave: case IpcId::TextDocumentRename: case IpcId::TextDocumentCompletion: case IpcId::TextDocumentSignatureHelp: case IpcId::TextDocumentDefinition: case IpcId::TextDocumentDocumentHighlight: case IpcId::TextDocumentHover: case IpcId::TextDocumentReferences: case IpcId::TextDocumentDocumentSymbol: case IpcId::TextDocumentDocumentLink: case IpcId::TextDocumentCodeAction: case IpcId::TextDocumentCodeLens: case IpcId::WorkspaceSymbol: case IpcId::CqueryFreshenIndex: case IpcId::CqueryTypeHierarchyTree: case IpcId::CqueryCallTreeInitial: case IpcId::CqueryCallTreeExpand: case IpcId::CqueryVars: case IpcId::CqueryCallers: case IpcId::CqueryBase: case IpcId::CqueryDerived: case IpcId::CqueryIndexFile: case IpcId::CqueryQueryDbWaitForIdleIndexer: { ipc->for_querydb.Enqueue(std::move(message)); break; } default: { LOG_S(ERROR) << "Unhandled IPC message " << IpcIdToString(message->method_id); exit(1); } } return WorkThread::Result::MoreWork; }); } void LaunchStdoutThread(std::unordered_map* request_times, MultiQueueWaiter* waiter, QueueManager* queue) { WorkThread::StartThread("stdout", [=]() { IpcManager* ipc = IpcManager::instance(); std::vector messages = ipc->for_stdout.DequeueAll(); if (messages.empty()) { waiter->Wait({&ipc->for_stdout}); return queue->HasWork() ? WorkThread::Result::MoreWork : WorkThread::Result::NoWork; } for (auto& message : messages) { if (ShouldDisplayIpcTiming(message.id)) { Timer time = (*request_times)[message.id]; time.ResetAndPrint("[e2e] Running " + std::string(IpcIdToString(message.id))); } if (g_log_stdin_stdout_to_stderr) { std::ostringstream sstream; sstream << "[COUT] |"; sstream << message.content; sstream << "|\n"; std::cerr << sstream.str(); std::cerr.flush(); } std::cout << message.content; std::cout.flush(); } return WorkThread::Result::MoreWork; }); } void LanguageServerMain(const std::string& bin_name, Config* config, MultiQueueWaiter* waiter) { QueueManager queue(waiter); std::unordered_map request_times; LaunchStdinLoop(config, &request_times); // We run a dedicated thread for writing to stdout because there can be an // unknown number of delays when output information. LaunchStdoutThread(&request_times, waiter, &queue); // Start querydb which takes over this thread. The querydb will launch // indexer threads as needed. RunQueryDbThread(bin_name, config, waiter, &queue); } //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// // MAIN //////////////////////////////////////////////////////////////////////// //////////////////////////////////////////////////////////////////////////////// int main(int argc, char** argv) { std::unordered_map options = ParseOptions(argc, argv); if (!HasOption(options, "--log-all-to-stderr")) loguru::g_stderr_verbosity = loguru::Verbosity_WARNING; loguru::g_flush_interval_ms = 0; loguru::init(argc, argv); MultiQueueWaiter waiter; IpcManager::CreateInstance(&waiter); // bool loop = true; // while (loop) // std::this_thread::sleep_for(std::chrono::milliseconds(10)); // std::this_thread::sleep_for(std::chrono::seconds(10)); PlatformInit(); IndexInit(); bool print_help = true; if (HasOption(options, "--log-file")) { loguru::add_file(options["--log-file"].c_str(), loguru::Truncate, loguru::Verbosity_MAX); } if (HasOption(options, "--log-stdin-stdout-to-stderr")) g_log_stdin_stdout_to_stderr = true; if (HasOption(options, "--clang-sanity-check")) { print_help = false; ClangSanityCheck(); } if (HasOption(options, "--test-unit")) { print_help = false; doctest::Context context; context.applyCommandLine(argc, argv); int res = context.run(); if (context.shouldExit()) return res; } if (HasOption(options, "--test-index")) { print_help = false; RunIndexTests(options["--test-index"]); #if defined(_WIN32) std::cerr << std::endl << "[Enter] to exit" << std::endl; std::cin.get(); #endif } if (HasOption(options, "--language-server")) { print_help = false; // std::cerr << "Running language server" << std::endl; auto config = MakeUnique(); LanguageServerMain(argv[0], config.get(), &waiter); return 0; } if (print_help) { std::cout << R"help(cquery is a low-latency C/C++/Objective-C language server. Command line options: --language-server Run as a language server. This implements the language server spec over STDIN and STDOUT. --test-unit Run unit tests. --test-index Run index tests. opt_filter_path can be used to specify which test to run. If not provided all tests are run. --log-stdin-stdout-to-stderr Print stdin and stdout messages to stderr. This is a aid for developing new language clients, as it makes it easier to figure out how the client is interacting with cquery. --log-file Emit diagnostic logging to the given path, which is taken literally, ie, it will be relative to the current working directory. --log-all-to-stderr Write all log messages to STDERR. --clang-sanity-check Run a simple index test. Verifies basic clang functionality. Needs to be executed from the cquery root checkout directory. --help Print this help information. Configuration: When opening up a directory, cquery will look for a compile_commands.json file emitted by your preferred build system. If not present, cquery will use a recursive directory listing instead. Command line flags can be provided by adding a file named `.cquery` in the top-level directory. Each line in that file is a separate argument. There are also a number of configuration options available when initializing the language server - your editor should have tooling to describe those options. See |Config| in this source code for a detailed list of all currently supported options. )help"; } return 0; }