From 4b0a3100c2dcb782b7ff2a38cc973473a9ee58b0 Mon Sep 17 00:00:00 2001 From: Marcel van der Veldt Date: Wed, 7 Oct 2020 17:38:17 +0200 Subject: [PATCH] silence logging --- music_assistant/helpers/process.py | 26 +++++++++----------------- 1 file changed, 9 insertions(+), 17 deletions(-) diff --git a/music_assistant/helpers/process.py b/music_assistant/helpers/process.py index 3fddfc2b..f65a3e28 100644 --- a/music_assistant/helpers/process.py +++ b/music_assistant/helpers/process.py @@ -45,7 +45,6 @@ class AsyncProcess(object): async def __aenter__(self) -> "AsyncProcess": """Enter context manager, start running the process in executor.""" - LOGGER.debug("[%s] Entered context manager", self._id) self.__proc_task = self.loop.run_in_executor(None, self.__run_proc) return self @@ -58,8 +57,6 @@ class AsyncProcess(object): exc_type, str(exc_value), ) - else: - LOGGER.debug("[%s] Context manager exit", self._id) self._exit = True # prevent a deadlock by clearing the queues @@ -71,18 +68,16 @@ class AsyncProcess(object): await self.__queue_out.get() self.__queue_out.task_done() await self.__proc_task - LOGGER.debug("[%s] Cleanup finished", self._id) + LOGGER.debug("[%s] Context manager closed", self._id) return True async def iterate_chunks(self) -> AsyncGenerator[bytes, None]: """Yield chunks from the output Queue. Generator.""" - LOGGER.debug("[%s] start reading from generator", self._id) while True: chunk = await self.read() yield chunk if not chunk or len(chunk) < self._chunksize: break - LOGGER.debug("[%s] finished reading from generator", self._id) async def read(self) -> bytes: """Read single chunk from the output Queue.""" @@ -117,7 +112,7 @@ class AsyncProcess(object): def __run_proc(self): """Run process in executor.""" try: - LOGGER.info( + LOGGER.debug( "[%s] Starting process with args: %s", self._id, str(self._process_args) ) proc = subprocess.Popen( @@ -142,9 +137,9 @@ class AsyncProcess(object): proc.wait() except Exception as exc: # pylint: disable=broad-except + LOGGER.warning("[%s] process exiting abormally: %s", self._id, str(exc)) LOGGER.exception(exc) finally: - LOGGER.error("[%s] process exiting", self._id) if proc.poll() is None: proc.terminate() proc.communicate() @@ -152,7 +147,6 @@ class AsyncProcess(object): def __write_stdin(self, _stdin): """Put chunks from queue to stdin.""" - LOGGER.debug("[%s] start write_stdin", self._id) try: while True: chunk = asyncio.run_coroutine_threadsafe( @@ -164,13 +158,12 @@ class AsyncProcess(object): break _stdin.write(chunk) except Exception as exc: # pylint: disable=broad-except - LOGGER.debug("[%s] write_stdin aborted (%s)", self._id, str(exc)) - else: - LOGGER.debug("[%s] write_stdin finished", self._id) + LOGGER.debug( + "[%s] write to stdin aborted with exception: %s", self._id, str(exc) + ) def __read_stdout(self, _stdout): """Put chunks from stdout to queue.""" - LOGGER.debug("[%s] start read_stdout", self._id) try: while True: chunk = _stdout.read(self._chunksize) @@ -178,11 +171,10 @@ class AsyncProcess(object): self.__queue_out.put(chunk), self.loop ).result() if not chunk or len(chunk) < self._chunksize: - LOGGER.debug("[%s] last chunk received on stdout", self._id) break # write empty chunk just in case asyncio.run_coroutine_threadsafe(self.__queue_out.put(b""), self.loop) except Exception as exc: # pylint: disable=broad-except - LOGGER.debug("[%s] read_stdout aborted (%s)", self._id, str(exc)) - else: - LOGGER.debug("[%s] read_stdout finished", self._id) + LOGGER.debug( + "[%s] read from stdout aborted with exception: %s", self._id, str(exc) + ) -- 2.34.1