run.py(448): self.define(self._stderr_path, StringSeries([])) --- modulename: string_series, funcname: __init__ string_series.py(51): if is_stringify_value(values): --- modulename: __init__, funcname: is_stringify_value __init__.py(123): return isinstance(var, StringifyValue) string_series.py(54): if not is_collection(values): --- modulename: __init__, funcname: is_collection __init__.py(137): return isinstance(var, (list, set, tuple)) string_series.py(57): self._truncated = any([len(value) > MAX_STRING_SERIES_VALUE_LENGTH for value in values]) --- modulename: string_series, funcname: string_series.py(57): self._truncated = any([len(value) > MAX_STRING_SERIES_VALUE_LENGTH for value in values]) string_series.py(58): self._values = [value[:MAX_STRING_SERIES_VALUE_LENGTH] for value in values] --- modulename: string_series, funcname: string_series.py(58): self._values = [value[:MAX_STRING_SERIES_VALUE_LENGTH] for value in values] string_series.py(60): if steps is None: string_series.py(61): self._steps = cycle([None]) string_series.py(66): if timestamps is None: string_series.py(67): self._timestamps = cycle([time.time()]) --- modulename: metadata_container, funcname: define metadata_container.py(358): with self._lock: metadata_container.py(359): old_attr = self.get_attribute(path) --- modulename: metadata_container, funcname: get_attribute metadata_container.py(374): with self._lock: metadata_container.py(375): return self._structure.get(parse_path(path)) --- modulename: paths, funcname: parse_path paths.py(26): return _remove_empty_paths(path.split("/")) --- modulename: paths, funcname: _remove_empty_paths paths.py(22): return list(filter(bool, paths)) --- modulename: container_structure, funcname: get container_structure.py(68): ref = self._structure container_structure.py(70): for index, part in enumerate(path): container_structure.py(71): if not isinstance(ref, self._node_type): container_structure.py(76): if part not in ref: --- modulename: _collections_abc, funcname: __contains__ _collections_abc.py(829): try: _collections_abc.py(830): self[key] --- modulename: namespace, funcname: __getitem__ namespace.py(64): return self._attributes[k] _collections_abc.py(834): return True container_structure.py(78): ref = ref[part] --- modulename: namespace, funcname: __getitem__ namespace.py(64): return self._attributes[k] container_structure.py(70): for index, part in enumerate(path): container_structure.py(71): if not isinstance(ref, self._node_type): container_structure.py(76): if part not in ref: --- modulename: _collections_abc, funcname: __contains__ _collections_abc.py(829): try: _collections_abc.py(830): self[key] --- modulename: namespace, funcname: __getitem__ namespace.py(64): return self._attributes[k] _collections_abc.py(834): return True container_structure.py(78): ref = ref[part] --- modulename: namespace, funcname: __getitem__ namespace.py(64): return self._attributes[k] container_structure.py(70): for index, part in enumerate(path): container_structure.py(71): if not isinstance(ref, self._node_type): container_structure.py(76): if part not in ref: --- modulename: _collections_abc, funcname: __contains__ _collections_abc.py(829): try: _collections_abc.py(830): self[key] --- modulename: namespace, funcname: __getitem__ namespace.py(64): return self._attributes[k] _collections_abc.py(831): except KeyError: _collections_abc.py(832): return False container_structure.py(77): return None metadata_container.py(374): with self._lock: metadata_container.py(360): if old_attr is not None: metadata_container.py(363): neptune_value = cast_value(value) --- modulename: type_casting, funcname: cast_value type_casting.py(57): from neptune.handler import Handler type_casting.py(59): from_stringify_value = False type_casting.py(60): if is_stringify_value(value): --- modulename: __init__, funcname: is_stringify_value __init__.py(123): return isinstance(var, StringifyValue) type_casting.py(63): if isinstance(value, Value): type_casting.py(64): return value metadata_container.py(364): if neptune_value is None: metadata_container.py(368): attr = ValueToAttributeVisitor(self, parse_path(path)).visit(neptune_value) --- modulename: paths, funcname: parse_path paths.py(26): return _remove_empty_paths(path.split("/")) --- modulename: paths, funcname: _remove_empty_paths paths.py(22): return list(filter(bool, paths)) --- modulename: value_to_attribute_visitor, funcname: __init__ value_to_attribute_visitor.py(63): self._container = container value_to_attribute_visitor.py(64): self._path = path --- modulename: value_visitor, funcname: visit value_visitor.py(50): return value.accept(self) --- modulename: string_series, funcname: accept string_series.py(81): return visitor.visit_string_series(self) --- modulename: value_to_attribute_visitor, funcname: visit_string_series value_to_attribute_visitor.py(94): return StringSeriesAttr(self._container, self._path) --- modulename: string_series, funcname: __init__ string_series.py(51): super().__init__(container, path) --- modulename: attribute, funcname: __init__ attribute.py(37): super().__init__() attribute.py(38): self._container = container attribute.py(39): self._path = path string_series.py(52): self._value_truncation_occurred = False metadata_container.py(369): self.set_attribute(path, attr) --- modulename: metadata_container, funcname: set_attribute metadata_container.py(378): with self._lock: metadata_container.py(379): return self._structure.set(parse_path(path), attribute) --- modulename: paths, funcname: parse_path paths.py(26): return _remove_empty_paths(path.split("/")) --- modulename: paths, funcname: _remove_empty_paths paths.py(22): return list(filter(bool, paths)) --- modulename: container_structure, funcname: set container_structure.py(83): ref = self._structure container_structure.py(84): location, attribute_name = path[:-1], path[-1] container_structure.py(86): for idx, part in enumerate(location): container_structure.py(87): if part not in ref: --- modulename: _collections_abc, funcname: __contains__ _collections_abc.py(829): try: _collections_abc.py(830): self[key] --- modulename: namespace, funcname: __getitem__ namespace.py(64): return self._attributes[k] _collections_abc.py(834): return True container_structure.py(89): ref = ref[part] --- modulename: namespace, funcname: __getitem__ namespace.py(64): return self._attributes[k] container_structure.py(90): if not isinstance(ref, self._node_type): container_structure.py(86): for idx, part in enumerate(location): container_structure.py(87): if part not in ref: --- modulename: _collections_abc, funcname: __contains__ _collections_abc.py(829): try: _collections_abc.py(830): self[key] --- modulename: namespace, funcname: __getitem__ namespace.py(64): return self._attributes[k] _collections_abc.py(834): return True container_structure.py(89): ref = ref[part] --- modulename: namespace, funcname: __getitem__ namespace.py(64): return self._attributes[k] container_structure.py(90): if not isinstance(ref, self._node_type): container_structure.py(86): for idx, part in enumerate(location): container_structure.py(96): if attribute_name in ref and isinstance(ref[attribute_name], self._node_type): --- modulename: _collections_abc, funcname: __contains__ _collections_abc.py(829): try: _collections_abc.py(830): self[key] --- modulename: namespace, funcname: __getitem__ namespace.py(64): return self._attributes[k] _collections_abc.py(831): except KeyError: _collections_abc.py(832): return False container_structure.py(102): ref[attribute_name] = attr --- modulename: namespace, funcname: __setitem__ namespace.py(58): self._attributes[k] = v metadata_container.py(378): with self._lock: metadata_container.py(370): attr.process_assignment(neptune_value, wait=wait) --- modulename: attribute, funcname: process_assignment attribute.py(63): if isinstance(value, ValueCopy): attribute.py(66): return self.assign(value, wait=wait) --- modulename: series, funcname: assign series.py(84): if not self._is_value_type(value): --- modulename: string_series, funcname: _is_value_type string_series.py(84): return isinstance(value, StringSeriesVal) series.py(86): clear_op = self._get_clear_operation() --- modulename: string_series, funcname: _get_clear_operation string_series.py(72): return ClearStringLog(self._path) --- modulename: operation, funcname: __init__ (3): series.py(87): config_op = self._get_config_operation_from_value(value) --- modulename: series, funcname: _get_config_operation_from_value series.py(66): return None series.py(88): with self._container.lock(): --- modulename: metadata_container, funcname: lock metadata_container.py(417): return self._lock series.py(89): if config_op: series.py(91): if not value.values: --- modulename: string_series, funcname: values string_series.py(85): return self._values series.py(92): self._enqueue_operation(clear_op, wait=wait) --- modulename: attribute, funcname: _enqueue_operation attribute.py(45): self._container._op_processor.enqueue_operation(operation, wait=wait) --- modulename: async_operation_processor, funcname: enqueue_operation async_operation_processor.py(103): if self._drop_operations: async_operation_processor.py(105): self._last_version = self._queue.put(op) --- modulename: disk_queue, funcname: put disk_queue.py(92): version = self._last_put_file.read_local() + 1 --- modulename: sync_offset_file, funcname: read_local sync_offset_file.py(44): return self._last disk_queue.py(93): _json = json.dumps(self._serialize(obj, version)) --- modulename: disk_queue, funcname: _serialize disk_queue.py(246): return {"obj": self._to_dict(obj), "version": version} --- modulename: async_operation_processor, funcname: async_operation_processor.py(68): to_dict=lambda x: x.to_dict(), --- modulename: operation, funcname: to_dict operation.py(67): return {"type": self.__class__.__name__, "path": self.path} --- modulename: __init__, funcname: dumps __init__.py(227): if (not skipkeys and ensure_ascii and __init__.py(228): check_circular and allow_nan and __init__.py(227): if (not skipkeys and ensure_ascii and __init__.py(228): check_circular and allow_nan and __init__.py(227): if (not skipkeys and ensure_ascii and __init__.py(229): cls is None and indent is None and separators is None and __init__.py(230): default is None and not sort_keys and not kw): __init__.py(231): return _default_encoder.encode(obj) --- modulename: encoder, funcname: encode encoder.py(191): if isinstance(o, str): encoder.py(199): chunks = self.iterencode(o, _one_shot=True) --- modulename: encoder, funcname: iterencode encoder.py(214): if self.check_circular: encoder.py(215): markers = {} encoder.py(218): if self.ensure_ascii: encoder.py(219): _encoder = encode_basestring_ascii encoder.py(223): def floatstr(o, allow_nan=self.allow_nan, encoder.py(224): _repr=float.__repr__, _inf=INFINITY, _neginf=-INFINITY): encoder.py(223): def floatstr(o, allow_nan=self.allow_nan, encoder.py(246): if (_one_shot and c_make_encoder is not None encoder.py(247): and self.indent is None): encoder.py(248): _iterencode = c_make_encoder( encoder.py(249): markers, self.default, _encoder, self.indent, encoder.py(250): self.key_separator, self.item_separator, self.sort_keys, encoder.py(251): self.skipkeys, self.allow_nan) encoder.py(248): _iterencode = c_make_encoder( encoder.py(257): return _iterencode(o, 0) encoder.py(200): if not isinstance(chunks, (list, tuple)): encoder.py(202): return ''.join(chunks) disk_queue.py(94): if self._file_size + len(_json) > self._max_file_size: disk_queue.py(100): self._writer.write(_json + "\n") disk_queue.py(101): self._last_put_file.write(version) --- modulename: sync_offset_file, funcname: write sync_offset_file.py(30): self._file.seek(0) --- modulename: codecs, funcname: reset codecs.py(328): IncrementalDecoder.reset(self) --- modulename: codecs, funcname: reset codecs.py(276): def reset(self): codecs.py(329): self.buffer = b"" --- modulename: codecs, funcname: reset codecs.py(203): def reset(self): sync_offset_file.py(31): self._file.write(str(offset)) --- modulename: codecs, funcname: reset codecs.py(328): IncrementalDecoder.reset(self) --- modulename: codecs, funcname: reset codecs.py(276): def reset(self): codecs.py(329): self.buffer = b"" sync_offset_file.py(32): self._file.truncate() sync_offset_file.py(33): self._file.flush() sync_offset_file.py(34): self._last = offset disk_queue.py(102): self._file_size += len(_json) + 1 disk_queue.py(103): return version async_operation_processor.py(106): if self._queue.size() > self._batch_size / 2: --- modulename: disk_queue, funcname: size disk_queue.py(223): return self._last_put_file.read_local() - self._last_ack_file.read_local() --- modulename: sync_offset_file, funcname: read_local sync_offset_file.py(44): return self._last --- modulename: sync_offset_file, funcname: read_local sync_offset_file.py(44): return self._last async_operation_processor.py(108): if wait: series.py(88): with self._container.lock(): metadata_container.py(371): return attr metadata_container.py(358): with self._lock: run.py(450): if self._with_id is None or self._source_files is not None: run.py(452): upload_source_code(source_files=self._source_files, run=self) --- modulename: source_code, funcname: upload_source_code source_code.py(43): entrypoint_filepath = get_path_executed_script() --- modulename: lib_programname, funcname: get_path_executed_script lib_programname.py(78): path_candidate = get_fullpath_from_main_file() --- modulename: lib_programname, funcname: get_fullpath_from_main_file lib_programname.py(100): if not hasattr(sys.modules["__main__"], "__file__"): lib_programname.py(103): arg_string = str(sys.modules["__main__"].__file__) lib_programname.py(104): valid_executable_path = get_valid_executable_path_or_empty_path(arg_string) --- modulename: lib_programname, funcname: get_valid_executable_path_or_empty_path lib_programname.py(136): arg_string = remove_doctest_and_docrunner_parameters(arg_string) --- modulename: lib_programname, funcname: remove_doctest_and_docrunner_parameters lib_programname.py(159): path = arg_string.split("::", 1)[0] lib_programname.py(160): return path lib_programname.py(137): arg_string = add_python_extension_if_not_there(arg_string) --- modulename: lib_programname, funcname: add_python_extension_if_not_there lib_programname.py(175): if not arg_string.endswith(".py"): lib_programname.py(177): return arg_string lib_programname.py(138): path = pathlib.Path(arg_string) --- modulename: pathlib, funcname: __new__ pathlib.py(958): if cls is Path: pathlib.py(959): cls = WindowsPath if os.name == 'nt' else PosixPath pathlib.py(960): self = cls._from_parts(args) --- modulename: pathlib, funcname: _from_parts pathlib.py(593): self = object.__new__(cls) pathlib.py(594): drv, root, parts = self._parse_args(args) --- modulename: pathlib, funcname: _parse_args pathlib.py(573): parts = [] pathlib.py(574): for a in args: pathlib.py(575): if isinstance(a, PurePath): pathlib.py(578): a = os.fspath(a) pathlib.py(579): if isinstance(a, str): pathlib.py(581): parts.append(str(a)) pathlib.py(574): for a in args: pathlib.py(587): return cls._flavour.parse_parts(parts) --- modulename: pathlib, funcname: parse_parts pathlib.py(57): parsed = [] pathlib.py(58): sep = self.sep pathlib.py(59): altsep = self.altsep pathlib.py(60): drv = root = '' pathlib.py(61): it = reversed(parts) pathlib.py(62): for part in it: pathlib.py(63): if not part: pathlib.py(65): if altsep: pathlib.py(67): drv, root, rel = self.splitroot(part) --- modulename: pathlib, funcname: splitroot pathlib.py(240): if part and part[0] == sep: pathlib.py(241): stripped_part = part.lstrip(sep) pathlib.py(247): if len(part) - len(stripped_part) == 2: pathlib.py(250): return '', sep, stripped_part pathlib.py(68): if sep in rel: pathlib.py(69): for x in reversed(rel.split(sep)): pathlib.py(70): if x and x != '.': pathlib.py(71): parsed.append(sys.intern(x)) pathlib.py(69): for x in reversed(rel.split(sep)): pathlib.py(70): if x and x != '.': pathlib.py(71): parsed.append(sys.intern(x)) pathlib.py(69): for x in reversed(rel.split(sep)): pathlib.py(70): if x and x != '.': pathlib.py(71): parsed.append(sys.intern(x)) pathlib.py(69): for x in reversed(rel.split(sep)): pathlib.py(70): if x and x != '.': pathlib.py(71): parsed.append(sys.intern(x)) pathlib.py(69): for x in reversed(rel.split(sep)): pathlib.py(70): if x and x != '.': pathlib.py(71): parsed.append(sys.intern(x)) pathlib.py(69): for x in reversed(rel.split(sep)): pathlib.py(70): if x and x != '.': pathlib.py(71): parsed.append(sys.intern(x)) pathlib.py(69): for x in reversed(rel.split(sep)): pathlib.py(70): if x and x != '.': pathlib.py(71): parsed.append(sys.intern(x)) pathlib.py(69): for x in reversed(rel.split(sep)): pathlib.py(70): if x and x != '.': pathlib.py(71): parsed.append(sys.intern(x)) pathlib.py(69): for x in reversed(rel.split(sep)): pathlib.py(70): if x and x != '.': pathlib.py(71): parsed.append(sys.intern(x)) pathlib.py(69): for x in reversed(rel.split(sep)): pathlib.py(75): if drv or root: pathlib.py(76): if not drv: pathlib.py(80): for part in it: pathlib.py(88): break pathlib.py(89): if drv or root: pathlib.py(90): parsed.append(drv + root) pathlib.py(91): parsed.reverse() pathlib.py(92): return drv, root, parsed pathlib.py(595): self._drv = drv pathlib.py(596): self._root = root pathlib.py(597): self._parts = parts pathlib.py(598): return self pathlib.py(961): if not self._flavour.is_supported: pathlib.py(964): return self lib_programname.py(139): try: lib_programname.py(140): if path.is_file(): --- modulename: pathlib, funcname: is_file pathlib.py(1321): try: pathlib.py(1322): return S_ISREG(self.stat().st_mode) --- modulename: pathlib, funcname: stat pathlib.py(1097): return self._accessor.stat(self, follow_symlinks=follow_symlinks) --- modulename: pathlib, funcname: __fspath__ pathlib.py(632): return str(self) --- modulename: pathlib, funcname: __str__ pathlib.py(624): try: pathlib.py(625): return self._str pathlib.py(626): except AttributeError: pathlib.py(627): self._str = self._format_parsed_parts(self._drv, self._root, pathlib.py(628): self._parts) or '.' pathlib.py(627): self._str = self._format_parsed_parts(self._drv, self._root, --- modulename: pathlib, funcname: _format_parsed_parts pathlib.py(610): if drv or root: pathlib.py(611): return drv + root + cls._flavour.join(parts[1:]) pathlib.py(629): return self._str lib_programname.py(141): path = path.resolve() # .resolve does not work on a non existing file in python 3.5 --- modulename: pathlib, funcname: resolve pathlib.py(1071): def check_eloop(e): pathlib.py(1076): try: pathlib.py(1077): s = self._accessor.realpath(self, strict=strict) --- modulename: posixpath, funcname: realpath posixpath.py(395): filename = os.fspath(filename) --- modulename: pathlib, funcname: __fspath__ pathlib.py(632): return str(self) --- modulename: pathlib, funcname: __str__ pathlib.py(624): try: pathlib.py(625): return self._str posixpath.py(396): path, ok = _joinrealpath(filename[:0], filename, strict, {}) --- modulename: posixpath, funcname: _joinrealpath posixpath.py(402): if isinstance(path, bytes): posixpath.py(407): sep = '/' posixpath.py(408): curdir = '.' posixpath.py(409): pardir = '..' posixpath.py(411): if isabs(rest): --- modulename: posixpath, funcname: isabs posixpath.py(62): s = os.fspath(s) posixpath.py(63): sep = _get_sep(s) --- modulename: posixpath, funcname: _get_sep posixpath.py(42): if isinstance(path, bytes): posixpath.py(45): return '/' posixpath.py(64): return s.startswith(sep) posixpath.py(412): rest = rest[1:] posixpath.py(413): path = sep posixpath.py(415): while rest: posixpath.py(416): name, _, rest = rest.partition(sep) posixpath.py(417): if not name or name == curdir: posixpath.py(420): if name == pardir: posixpath.py(429): newpath = join(path, name) --- modulename: posixpath, funcname: join posixpath.py(76): a = os.fspath(a) posixpath.py(77): sep = _get_sep(a) --- modulename: posixpath, funcname: _get_sep posixpath.py(42): if isinstance(path, bytes): posixpath.py(45): return '/' posixpath.py(78): path = a posixpath.py(79): try: posixpath.py(80): if not p: posixpath.py(82): for b in map(os.fspath, p): posixpath.py(83): if b.startswith(sep): posixpath.py(85): elif not path or path.endswith(sep): posixpath.py(86): path += b posixpath.py(82): for b in map(os.fspath, p): posixpath.py(92): return path posixpath.py(430): try: posixpath.py(431): st = os.lstat(newpath) posixpath.py(437): is_link = stat.S_ISLNK(st.st_mode) posixpath.py(438): if not is_link: posixpath.py(439): path = newpath posixpath.py(440): continue posixpath.py(415): while rest: posixpath.py(416): name, _, rest = rest.partition(sep) posixpath.py(417): if not name or name == curdir: posixpath.py(420): if name == pardir: posixpath.py(429): newpath = join(path, name) --- modulename: posixpath, funcname: join posixpath.py(76): a = os.fspath(a) posixpath.py(77): sep = _get_sep(a) --- modulename: posixpath, funcname: _get_sep posixpath.py(42): if isinstance(path, bytes): posixpath.py(45): return '/' posixpath.py(78): path = a posixpath.py(79): try: posixpath.py(80): if not p: posixpath.py(82): for b in map(os.fspath, p): posixpath.py(83): if b.startswith(sep): posixpath.py(85): elif not path or path.endswith(sep): posixpath.py(88): path += sep + b posixpath.py(82): for b in map(os.fspath, p): posixpath.py(92): return path posixpath.py(430): try: posixpath.py(431): st = os.lstat(newpath) posixpath.py(437): is_link = stat.S_ISLNK(st.st_mode) posixpath.py(438): if not is_link: posixpath.py(439): path = newpath posixpath.py(440): continue posixpath.py(415): while rest: posixpath.py(416): name, _, rest = rest.partition(sep) posixpath.py(417): if not name or name == curdir: posixpath.py(420): if name == pardir: posixpath.py(429): newpath = join(path, name) --- modulename: posixpath, funcname: join posixpath.py(76): a = os.fspath(a) posixpath.py(77): sep = _get_sep(a) --- modulename: posixpath, funcname: _get_sep posixpath.py(42): if isinstance(path, bytes): posixpath.py(45): return '/' posixpath.py(78): path = a posixpath.py(79): try: posixpath.py(80): if not p: posixpath.py(82): for b in map(os.fspath, p): posixpath.py(83): if b.startswith(sep): posixpath.py(85): elif not path or path.endswith(sep): posixpath.py(88): path += sep + b posixpath.py(82): for b in map(os.fspath, p): posixpath.py(92): return path posixpath.py(430): try: posixpath.py(431): st = os.lstat(newpath) posixpath.py(437): is_link = stat.S_ISLNK(st.st_mode) posixpath.py(438): if not is_link: posixpath.py(439): path = newpath posixpath.py(440): continue posixpath.py(415): while rest: posixpath.py(416): name, _, rest = rest.partition(sep) posixpath.py(417): if not name or name == curdir: posixpath.py(420): if name == pardir: posixpath.py(429): newpath = join(path, name) --- modulename: posixpath, funcname: join posixpath.py(76): a = os.fspath(a) posixpath.py(77): sep = _get_sep(a) --- modulename: posixpath, funcname: _get_sep posixpath.py(42): if isinstance(path, bytes): posixpath.py(45): return '/' posixpath.py(78): path = a posixpath.py(79): try: posixpath.py(80): if not p: posixpath.py(82): for b in map(os.fspath, p): posixpath.py(83): if b.startswith(sep): posixpath.py(85): elif not path or path.endswith(sep): posixpath.py(88): path += sep + b posixpath.py(82): for b in map(os.fspath, p): posixpath.py(92): return path posixpath.py(430): try: posixpath.py(431): st = os.lstat(newpath) posixpath.py(437): is_link = stat.S_ISLNK(st.st_mode) posixpath.py(438): if not is_link: posixpath.py(439): path = newpath posixpath.py(440): continue posixpath.py(415): while rest: posixpath.py(416): name, _, rest = rest.partition(sep) posixpath.py(417): if not name or name == curdir: posixpath.py(420): if name == pardir: posixpath.py(429): newpath = join(path, name) --- modulename: posixpath, funcname: join posixpath.py(76): a = os.fspath(a) posixpath.py(77): sep = _get_sep(a) --- modulename: posixpath, funcname: _get_sep posixpath.py(42): if isinstance(path, bytes): posixpath.py(45): return '/' posixpath.py(78): path = a posixpath.py(79): try: posixpath.py(80): if not p: posixpath.py(82): for b in map(os.fspath, p): posixpath.py(83): if b.startswith(sep): posixpath.py(85): elif not path or path.endswith(sep): posixpath.py(88): path += sep + b posixpath.py(82): for b in map(os.fspath, p): posixpath.py(92): return path posixpath.py(430): try: posixpath.py(431): st = os.lstat(newpath) posixpath.py(437): is_link = stat.S_ISLNK(st.st_mode) posixpath.py(438): if not is_link: posixpath.py(439): path = newpath posixpath.py(440): continue posixpath.py(415): while rest: posixpath.py(416): name, _, rest = rest.partition(sep) posixpath.py(417): if not name or name == curdir: posixpath.py(420): if name == pardir: posixpath.py(429): newpath = join(path, name) --- modulename: posixpath, funcname: join posixpath.py(76): a = os.fspath(a) posixpath.py(77): sep = _get_sep(a) --- modulename: posixpath, funcname: _get_sep posixpath.py(42): if isinstance(path, bytes): posixpath.py(45): return '/' posixpath.py(78): path = a posixpath.py(79): try: posixpath.py(80): if not p: posixpath.py(82): for b in map(os.fspath, p): posixpath.py(83): if b.startswith(sep): posixpath.py(85): elif not path or path.endswith(sep): posixpath.py(88): path += sep + b posixpath.py(82): for b in map(os.fspath, p): posixpath.py(92): return path posixpath.py(430): try: posixpath.py(431): st = os.lstat(newpath) posixpath.py(437): is_link = stat.S_ISLNK(st.st_mode) posixpath.py(438): if not is_link: posixpath.py(439): path = newpath posixpath.py(440): continue posixpath.py(415): while rest: posixpath.py(416): name, _, rest = rest.partition(sep) posixpath.py(417): if not name or name == curdir: posixpath.py(420): if name == pardir: posixpath.py(429): newpath = join(path, name) --- modulename: posixpath, funcname: join posixpath.py(76): a = os.fspath(a) posixpath.py(77): sep = _get_sep(a) --- modulename: posixpath, funcname: _get_sep posixpath.py(42): if isinstance(path, bytes): posixpath.py(45): return '/' posixpath.py(78): path = a posixpath.py(79): try: posixpath.py(80): if not p: posixpath.py(82): for b in map(os.fspath, p): posixpath.py(83): if b.startswith(sep): posixpath.py(85): elif not path or path.endswith(sep): posixpath.py(88): path += sep + b posixpath.py(82): for b in map(os.fspath, p): posixpath.py(92): return path posixpath.py(430): try: posixpath.py(431): st = os.lstat(newpath) posixpath.py(437): is_link = stat.S_ISLNK(st.st_mode) posixpath.py(438): if not is_link: posixpath.py(439): path = newpath posixpath.py(440): continue posixpath.py(415): while rest: posixpath.py(416): name, _, rest = rest.partition(sep) posixpath.py(417): if not name or name == curdir: posixpath.py(420): if name == pardir: posixpath.py(429): newpath = join(path, name) --- modulename: posixpath, funcname: join posixpath.py(76): a = os.fspath(a) posixpath.py(77): sep = _get_sep(a) --- modulename: posixpath, funcname: _get_sep posixpath.py(42): if isinstance(path, bytes): posixpath.py(45): return '/' posixpath.py(78): path = a posixpath.py(79): try: posixpath.py(80): if not p: posixpath.py(82): for b in map(os.fspath, p): posixpath.py(83): if b.startswith(sep): posixpath.py(85): elif not path or path.endswith(sep): posixpath.py(88): path += sep + b posixpath.py(82): for b in map(os.fspath, p): posixpath.py(92): return path posixpath.py(430): try: posixpath.py(431): st = os.lstat(newpath) posixpath.py(437): is_link = stat.S_ISLNK(st.st_mode) posixpath.py(438): if not is_link: posixpath.py(439): path = newpath posixpath.py(440): continue posixpath.py(415): while rest: posixpath.py(416): name, _, rest = rest.partition(sep) posixpath.py(417): if not name or name == curdir: posixpath.py(420): if name == pardir: posixpath.py(429): newpath = join(path, name) --- modulename: posixpath, funcname: join posixpath.py(76): a = os.fspath(a) posixpath.py(77): sep = _get_sep(a) --- modulename: posixpath, funcname: _get_sep posixpath.py(42): if isinstance(path, bytes): posixpath.py(45): return '/' posixpath.py(78): path = a posixpath.py(79): try: posixpath.py(80): if not p: posixpath.py(82): for b in map(os.fspath, p): posixpath.py(83): if b.startswith(sep): posixpath.py(85): elif not path or path.endswith(sep): posixpath.py(88): path += sep + b posixpath.py(82): for b in map(os.fspath, p): posixpath.py(92): return path posixpath.py(430): try: posixpath.py(431): st = os.lstat(newpath) posixpath.py(437): is_link = stat.S_ISLNK(st.st_mode) posixpath.py(438): if not is_link: posixpath.py(439): path = newpath posixpath.py(440): continue posixpath.py(415): while rest: posixpath.py(461): return path, True posixpath.py(397): return abspath(path) --- modulename: posixpath, funcname: abspath posixpath.py(379): path = os.fspath(path) posixpath.py(380): if not isabs(path): --- modulename: posixpath, funcname: isabs posixpath.py(62): s = os.fspath(s) posixpath.py(63): sep = _get_sep(s) --- modulename: posixpath, funcname: _get_sep posixpath.py(42): if isinstance(path, bytes): posixpath.py(45): return '/' posixpath.py(64): return s.startswith(sep) posixpath.py(386): return normpath(path) --- modulename: posixpath, funcname: normpath posixpath.py(340): path = os.fspath(path) posixpath.py(341): if isinstance(path, bytes): posixpath.py(347): sep = '/' posixpath.py(348): empty = '' posixpath.py(349): dot = '.' posixpath.py(350): dotdot = '..' posixpath.py(351): if path == empty: posixpath.py(353): initial_slashes = path.startswith(sep) posixpath.py(357): if (initial_slashes and posixpath.py(358): path.startswith(sep*2) and not path.startswith(sep*3)): posixpath.py(357): if (initial_slashes and posixpath.py(360): comps = path.split(sep) posixpath.py(361): new_comps = [] posixpath.py(362): for comp in comps: posixpath.py(363): if comp in (empty, dot): posixpath.py(364): continue posixpath.py(362): for comp in comps: posixpath.py(363): if comp in (empty, dot): posixpath.py(365): if (comp != dotdot or (not initial_slashes and not new_comps) or posixpath.py(367): new_comps.append(comp) posixpath.py(362): for comp in comps: posixpath.py(363): if comp in (empty, dot): posixpath.py(365): if (comp != dotdot or (not initial_slashes and not new_comps) or posixpath.py(367): new_comps.append(comp) posixpath.py(362): for comp in comps: posixpath.py(363): if comp in (empty, dot): posixpath.py(365): if (comp != dotdot or (not initial_slashes and not new_comps) or posixpath.py(367): new_comps.append(comp) posixpath.py(362): for comp in comps: posixpath.py(363): if comp in (empty, dot): posixpath.py(365): if (comp != dotdot or (not initial_slashes and not new_comps) or posixpath.py(367): new_comps.append(comp) posixpath.py(362): for comp in comps: posixpath.py(363): if comp in (empty, dot): posixpath.py(365): if (comp != dotdot or (not initial_slashes and not new_comps) or posixpath.py(367): new_comps.append(comp) posixpath.py(362): for comp in comps: posixpath.py(363): if comp in (empty, dot): posixpath.py(365): if (comp != dotdot or (not initial_slashes and not new_comps) or posixpath.py(367): new_comps.append(comp) posixpath.py(362): for comp in comps: posixpath.py(363): if comp in (empty, dot): posixpath.py(365): if (comp != dotdot or (not initial_slashes and not new_comps) or posixpath.py(367): new_comps.append(comp) posixpath.py(362): for comp in comps: posixpath.py(363): if comp in (empty, dot): posixpath.py(365): if (comp != dotdot or (not initial_slashes and not new_comps) or posixpath.py(367): new_comps.append(comp) posixpath.py(362): for comp in comps: posixpath.py(363): if comp in (empty, dot): posixpath.py(365): if (comp != dotdot or (not initial_slashes and not new_comps) or posixpath.py(367): new_comps.append(comp) posixpath.py(362): for comp in comps: posixpath.py(370): comps = new_comps posixpath.py(371): path = sep.join(comps) posixpath.py(372): if initial_slashes: posixpath.py(373): path = sep*initial_slashes + path posixpath.py(374): return path or dot pathlib.py(1081): p = self._from_parts((s,)) --- modulename: pathlib, funcname: _from_parts pathlib.py(593): self = object.__new__(cls) pathlib.py(594): drv, root, parts = self._parse_args(args) --- modulename: pathlib, funcname: _parse_args pathlib.py(573): parts = [] pathlib.py(574): for a in args: pathlib.py(575): if isinstance(a, PurePath): pathlib.py(578): a = os.fspath(a) pathlib.py(579): if isinstance(a, str): pathlib.py(581): parts.append(str(a)) pathlib.py(574): for a in args: pathlib.py(587): return cls._flavour.parse_parts(parts) --- modulename: pathlib, funcname: parse_parts pathlib.py(57): parsed = [] pathlib.py(58): sep = self.sep pathlib.py(59): altsep = self.altsep pathlib.py(60): drv = root = '' pathlib.py(61): it = reversed(parts) pathlib.py(62): for part in it: pathlib.py(63): if not part: pathlib.py(65): if altsep: pathlib.py(67): drv, root, rel = self.splitroot(part) --- modulename: pathlib, funcname: splitroot pathlib.py(240): if part and part[0] == sep: pathlib.py(241): stripped_part = part.lstrip(sep) pathlib.py(247): if len(part) - len(stripped_part) == 2: pathlib.py(250): return '', sep, stripped_part pathlib.py(68): if sep in rel: pathlib.py(69): for x in reversed(rel.split(sep)): pathlib.py(70): if x and x != '.': pathlib.py(71): parsed.append(sys.intern(x)) pathlib.py(69): for x in reversed(rel.split(sep)): pathlib.py(70): if x and x != '.': pathlib.py(71): parsed.append(sys.intern(x)) pathlib.py(69): for x in reversed(rel.split(sep)): pathlib.py(70): if x and x != '.': pathlib.py(71): parsed.append(sys.intern(x)) pathlib.py(69): for x in reversed(rel.split(sep)): pathlib.py(70): if x and x != '.': pathlib.py(71): parsed.append(sys.intern(x)) pathlib.py(69): for x in reversed(rel.split(sep)): pathlib.py(70): if x and x != '.': pathlib.py(71): parsed.append(sys.intern(x)) pathlib.py(69): for x in reversed(rel.split(sep)): pathlib.py(70): if x and x != '.': pathlib.py(71): parsed.append(sys.intern(x)) pathlib.py(69): for x in reversed(rel.split(sep)): pathlib.py(70): if x and x != '.': pathlib.py(71): parsed.append(sys.intern(x)) pathlib.py(69): for x in reversed(rel.split(sep)): pathlib.py(70): if x and x != '.': pathlib.py(71): parsed.append(sys.intern(x)) pathlib.py(69): for x in reversed(rel.split(sep)): pathlib.py(70): if x and x != '.': pathlib.py(71): parsed.append(sys.intern(x)) pathlib.py(69): for x in reversed(rel.split(sep)): pathlib.py(75): if drv or root: pathlib.py(76): if not drv: pathlib.py(80): for part in it: pathlib.py(88): break pathlib.py(89): if drv or root: pathlib.py(90): parsed.append(drv + root) pathlib.py(91): parsed.reverse() pathlib.py(92): return drv, root, parsed pathlib.py(595): self._drv = drv pathlib.py(596): self._root = root pathlib.py(597): self._parts = parts pathlib.py(598): return self pathlib.py(1085): if not strict: pathlib.py(1086): try: pathlib.py(1087): p.stat() --- modulename: pathlib, funcname: stat pathlib.py(1097): return self._accessor.stat(self, follow_symlinks=follow_symlinks) --- modulename: pathlib, funcname: __fspath__ pathlib.py(632): return str(self) --- modulename: pathlib, funcname: __str__ pathlib.py(624): try: pathlib.py(625): return self._str pathlib.py(626): except AttributeError: pathlib.py(627): self._str = self._format_parsed_parts(self._drv, self._root, pathlib.py(628): self._parts) or '.' pathlib.py(627): self._str = self._format_parsed_parts(self._drv, self._root, --- modulename: pathlib, funcname: _format_parsed_parts pathlib.py(610): if drv or root: pathlib.py(611): return drv + root + cls._flavour.join(parts[1:]) pathlib.py(629): return self._str pathlib.py(1090): return p lib_programname.py(142): return path lib_programname.py(105): return valid_executable_path lib_programname.py(79): if path_candidate != empty_path: --- modulename: pathlib, funcname: __eq__ pathlib.py(664): if not isinstance(other, PurePath): pathlib.py(666): return self._cparts == other._cparts and self._flavour is other._flavour --- modulename: pathlib, funcname: _cparts pathlib.py(657): try: pathlib.py(658): return self._cached_cparts pathlib.py(659): except AttributeError: pathlib.py(660): self._cached_cparts = self._flavour.casefold_parts(self._parts) --- modulename: pathlib, funcname: casefold_parts pathlib.py(258): return parts pathlib.py(661): return self._cached_cparts --- modulename: pathlib, funcname: _cparts pathlib.py(657): try: pathlib.py(658): return self._cached_cparts lib_programname.py(80): return path_candidate source_code.py(45): if not is_ipython() and entrypoint_filepath != empty_path and os.path.isfile(entrypoint_filepath): --- modulename: utils, funcname: is_ipython utils.py(218): try: utils.py(219): import IPython utils.py(221): ipython = IPython.core.getipython.get_ipython() --- modulename: getipython, funcname: get_ipython getipython.py(22): from IPython.core.interactiveshell import InteractiveShell getipython.py(23): if InteractiveShell.initialized(): --- modulename: configurable, funcname: initialized configurable.py(568): return hasattr(cls, "_instance") and cls._instance is not None utils.py(222): return ipython is not None --- modulename: pathlib, funcname: __eq__ pathlib.py(664): if not isinstance(other, PurePath): pathlib.py(666): return self._cparts == other._cparts and self._flavour is other._flavour --- modulename: pathlib, funcname: _cparts pathlib.py(657): try: pathlib.py(658): return self._cached_cparts --- modulename: pathlib, funcname: _cparts pathlib.py(657): try: pathlib.py(658): return self._cached_cparts --- modulename: genericpath, funcname: isfile genericpath.py(29): try: genericpath.py(30): st = os.stat(path) --- modulename: pathlib, funcname: __fspath__ pathlib.py(632): return str(self) --- modulename: pathlib, funcname: __str__ pathlib.py(624): try: pathlib.py(625): return self._str genericpath.py(33): return stat.S_ISREG(st.st_mode) source_code.py(46): if source_files is None: source_code.py(47): entrypoint = os.path.basename(entrypoint_filepath) --- modulename: posixpath, funcname: basename posixpath.py(142): p = os.fspath(p) --- modulename: pathlib, funcname: __fspath__ pathlib.py(632): return str(self) --- modulename: pathlib, funcname: __str__ pathlib.py(624): try: pathlib.py(625): return self._str posixpath.py(143): sep = _get_sep(p) --- modulename: posixpath, funcname: _get_sep posixpath.py(42): if isinstance(path, bytes): posixpath.py(45): return '/' posixpath.py(144): i = p.rfind(sep) + 1 posixpath.py(145): return p[i:] source_code.py(48): source_files = str(entrypoint_filepath) --- modulename: pathlib, funcname: __str__ pathlib.py(624): try: pathlib.py(625): return self._str source_code.py(60): run[attr_consts.SOURCE_CODE_ENTRYPOINT_ATTRIBUTE_PATH] = entrypoint --- modulename: metadata_container, funcname: inner_fun metadata_container.py(83): self._raise_if_stopped() --- modulename: run, funcname: _raise_if_stopped run.py(459): if self._state == ContainerState.STOPPED: metadata_container.py(84): return fun(self, *args, **kwargs) --- modulename: metadata_container, funcname: __setitem__ metadata_container.py(188): self.__getitem__(key).assign(value) --- modulename: metadata_container, funcname: inner_fun metadata_container.py(83): self._raise_if_stopped() --- modulename: run, funcname: _raise_if_stopped run.py(459): if self._state == ContainerState.STOPPED: metadata_container.py(84): return fun(self, *args, **kwargs) --- modulename: metadata_container, funcname: __getitem__ metadata_container.py(184): return Handler(self, path) --- modulename: handler, funcname: __init__ handler.py(101): super().__init__() handler.py(102): self._container = container handler.py(103): self._path = path --- modulename: handler, funcname: inner_fun handler.py(85): validate_path_not_protected(self._path, self) --- modulename: handler, funcname: validate_path_not_protected handler.py(77): path_protection_exception = handler._PROTECTED_PATHS.get(target_path) handler.py(78): if path_protection_exception: handler.py(86): return fun(self, *args, **kwargs) --- modulename: handler, funcname: assign handler.py(208): with self._container.lock(): --- modulename: metadata_container, funcname: lock metadata_container.py(417): return self._lock handler.py(209): attr = self._container.get_attribute(self._path) --- modulename: metadata_container, funcname: get_attribute metadata_container.py(374): with self._lock: metadata_container.py(375): return self._structure.get(parse_path(path)) --- modulename: paths, funcname: parse_path paths.py(26): return _remove_empty_paths(path.split("/")) --- modulename: paths, funcname: _remove_empty_paths paths.py(22): return list(filter(bool, paths)) --- modulename: container_structure, funcname: get container_structure.py(68): ref = self._structure container_structure.py(70): for index, part in enumerate(path): container_structure.py(71): if not isinstance(ref, self._node_type): container_structure.py(76): if part not in ref: --- modulename: _collections_abc, funcname: __contains__ _collections_abc.py(829): try: _collections_abc.py(830): self[key] --- modulename: namespace, funcname: __getitem__ namespace.py(64): return self._attributes[k] _collections_abc.py(831): except KeyError: _collections_abc.py(832): return False container_structure.py(77): return None metadata_container.py(374): with self._lock: handler.py(210): if attr is None: handler.py(211): self._container.define(self._path, value) --- modulename: metadata_container, funcname: define metadata_container.py(358): with self._lock: metadata_container.py(359): old_attr = self.get_attribute(path) --- modulename: metadata_container, funcname: get_attribute metadata_container.py(374): with self._lock: metadata_container.py(375): return self._structure.get(parse_path(path)) --- modulename: paths, funcname: parse_path paths.py(26): return _remove_empty_paths(path.split("/")) --- modulename: paths, funcname: _remove_empty_paths paths.py(22): return list(filter(bool, paths)) --- modulename: container_structure, funcname: get container_structure.py(68): ref = self._structure container_structure.py(70): for index, part in enumerate(path): container_structure.py(71): if not isinstance(ref, self._node_type): container_structure.py(76): if part not in ref: --- modulename: _collections_abc, funcname: __contains__ _collections_abc.py(829): try: _collections_abc.py(830): self[key] --- modulename: namespace, funcname: __getitem__ namespace.py(64): return self._attributes[k] _collections_abc.py(831): except KeyError: _collections_abc.py(832): return False container_structure.py(77): return None metadata_container.py(374): with self._lock: metadata_container.py(360): if old_attr is not None: metadata_container.py(363): neptune_value = cast_value(value) --- modulename: type_casting, funcname: cast_value type_casting.py(57): from neptune.handler import Handler type_casting.py(59): from_stringify_value = False type_casting.py(60): if is_stringify_value(value): --- modulename: __init__, funcname: is_stringify_value __init__.py(123): return isinstance(var, StringifyValue) type_casting.py(63): if isinstance(value, Value): type_casting.py(65): elif isinstance(value, Handler): --- modulename: abc, funcname: __instancecheck__ abc.py(119): return _abc_instancecheck(cls, instance) type_casting.py(67): elif isinstance(value, argparse.Namespace): type_casting.py(69): elif File.is_convertable_to_image(value): --- modulename: file, funcname: is_convertable_to_image file.py(315): convertable_to_img_predicates = (is_pil_image, is_matplotlib_figure) file.py(316): return any(predicate(value) for predicate in convertable_to_img_predicates) --- modulename: file, funcname: file.py(316): return any(predicate(value) for predicate in convertable_to_img_predicates) --- modulename: images, funcname: is_pil_image images.py(261): return PILImage is not None and isinstance(image, PILImage) --- modulename: file, funcname: file.py(316): return any(predicate(value) for predicate in convertable_to_img_predicates) --- modulename: images, funcname: is_matplotlib_figure images.py(265): return image.__class__.__module__.startswith("matplotlib.") and image.__class__.__name__ == "Figure" --- modulename: file, funcname: file.py(316): return any(predicate(value) for predicate in convertable_to_img_predicates) type_casting.py(71): elif File.is_convertable_to_html(value): --- modulename: file, funcname: is_convertable_to_html file.py(320): convertable_to_html_predicates = (is_altair_chart, is_bokeh_figure, is_plotly_figure) file.py(321): return any(predicate(value) for predicate in convertable_to_html_predicates) --- modulename: file, funcname: file.py(321): return any(predicate(value) for predicate in convertable_to_html_predicates) --- modulename: images, funcname: is_altair_chart images.py(273): return chart.__class__.__module__.startswith("altair.") and "Chart" in chart.__class__.__name__ --- modulename: file, funcname: file.py(321): return any(predicate(value) for predicate in convertable_to_html_predicates) --- modulename: images, funcname: is_bokeh_figure images.py(277): return chart.__class__.__module__.startswith("bokeh.") and chart.__class__.__name__.lower() == "figure" --- modulename: file, funcname: file.py(321): return any(predicate(value) for predicate in convertable_to_html_predicates) --- modulename: images, funcname: is_plotly_figure images.py(269): return chart.__class__.__module__.startswith("plotly.") and chart.__class__.__name__ == "Figure" --- modulename: file, funcname: file.py(321): return any(predicate(value) for predicate in convertable_to_html_predicates) type_casting.py(73): elif is_bool(value): --- modulename: __init__, funcname: is_bool __init__.py(87): return isinstance(var, bool) type_casting.py(75): elif is_int(value): --- modulename: __init__, funcname: is_int __init__.py(91): return isinstance(var, int) type_casting.py(77): elif is_float(value): --- modulename: __init__, funcname: is_float __init__.py(95): return isinstance(var, (float, int)) type_casting.py(79): elif is_string(value): --- modulename: __init__, funcname: is_string __init__.py(99): return isinstance(var, str) type_casting.py(80): return String(value) --- modulename: string, funcname: __init__ string.py(45): verify_type("value", value, (str, type(None), StringifyValue)) --- modulename: __init__, funcname: verify_type __init__.py(66): try: __init__.py(67): if isinstance(expected_type, tuple): __init__.py(68): type_name = " or ".join(get_type_name(t) for t in expected_type) --- modulename: __init__, funcname: __init__.py(68): type_name = " or ".join(get_type_name(t) for t in expected_type) --- modulename: __init__, funcname: get_type_name __init__.py(127): return _type.__name__ if hasattr(_type, "__name__") else str(_type) --- modulename: __init__, funcname: __init__.py(68): type_name = " or ".join(get_type_name(t) for t in expected_type) --- modulename: __init__, funcname: get_type_name __init__.py(127): return _type.__name__ if hasattr(_type, "__name__") else str(_type) --- modulename: __init__, funcname: __init__.py(68): type_name = " or ".join(get_type_name(t) for t in expected_type) --- modulename: __init__, funcname: get_type_name __init__.py(127): return _type.__name__ if hasattr(_type, "__name__") else str(_type) --- modulename: __init__, funcname: __init__.py(68): type_name = " or ".join(get_type_name(t) for t in expected_type) __init__.py(75): if not isinstance(var, expected_type): __init__.py(78): if isinstance(var, IOBase) and not hasattr(var, "read"): --- modulename: abc, funcname: __instancecheck__ abc.py(119): return _abc_instancecheck(cls, instance) string.py(47): self.value = str(value.value) if is_stringify_value(value) else value --- modulename: __init__, funcname: is_stringify_value __init__.py(123): return isinstance(var, StringifyValue) metadata_container.py(364): if neptune_value is None: metadata_container.py(368): attr = ValueToAttributeVisitor(self, parse_path(path)).visit(neptune_value) --- modulename: paths, funcname: parse_path paths.py(26): return _remove_empty_paths(path.split("/")) --- modulename: paths, funcname: _remove_empty_paths paths.py(22): return list(filter(bool, paths)) --- modulename: value_to_attribute_visitor, funcname: __init__ value_to_attribute_visitor.py(63): self._container = container value_to_attribute_visitor.py(64): self._path = path --- modulename: value_visitor, funcname: visit value_visitor.py(50): return value.accept(self) --- modulename: string, funcname: accept string.py(50): return visitor.visit_string(self) --- modulename: value_to_attribute_visitor, funcname: visit_string value_to_attribute_visitor.py(76): return StringAttr(self._container, self._path) --- modulename: string, funcname: __init__ string.py(37): super().__init__(container, path) --- modulename: attribute, funcname: __init__ attribute.py(37): super().__init__() attribute.py(38): self._container = container attribute.py(39): self._path = path string.py(38): self._value_truncation_occurred = False metadata_container.py(369): self.set_attribute(path, attr) --- modulename: metadata_container, funcname: set_attribute metadata_container.py(378): with self._lock: metadata_container.py(379): return self._structure.set(parse_path(path), attribute) --- modulename: paths, funcname: parse_path paths.py(26): return _remove_empty_paths(path.split("/")) --- modulename: paths, funcname: _remove_empty_paths paths.py(22): return list(filter(bool, paths)) --- modulename: container_structure, funcname: set container_structure.py(83): ref = self._structure container_structure.py(84): location, attribute_name = path[:-1], path[-1] container_structure.py(86): for idx, part in enumerate(location): container_structure.py(87): if part not in ref: --- modulename: _collections_abc, funcname: __contains__ _collections_abc.py(829): try: _collections_abc.py(830): self[key] --- modulename: namespace, funcname: __getitem__ namespace.py(64): return self._attributes[k] _collections_abc.py(831): except KeyError: _collections_abc.py(832): return False container_structure.py(88): ref[part] = self._node_factory(location[: idx + 1]) --- modulename: namespace, funcname: __call__ namespace.py(129): return Namespace(self._run, path) --- modulename: namespace, funcname: __init__ namespace.py(53): Attribute.__init__(self, container, path) --- modulename: attribute, funcname: __init__ attribute.py(37): super().__init__() attribute.py(38): self._container = container attribute.py(39): self._path = path namespace.py(54): self._attributes = {} namespace.py(55): self._str_path = path_to_str(path) --- modulename: paths, funcname: path_to_str paths.py(30): return "/".join(_remove_empty_paths(path)) --- modulename: paths, funcname: _remove_empty_paths paths.py(22): return list(filter(bool, paths)) --- modulename: namespace, funcname: __setitem__ namespace.py(58): self._attributes[k] = v container_structure.py(89): ref = ref[part] --- modulename: namespace, funcname: __getitem__ namespace.py(64): return self._attributes[k] container_structure.py(90): if not isinstance(ref, self._node_type): container_structure.py(86): for idx, part in enumerate(location): container_structure.py(96): if attribute_name in ref and isinstance(ref[attribute_name], self._node_type): --- modulename: _collections_abc, funcname: __contains__ _collections_abc.py(829): try: _collections_abc.py(830): self[key] --- modulename: namespace, funcname: __getitem__ namespace.py(64): return self._attributes[k] _collections_abc.py(831): except KeyError: _collections_abc.py(832): return False container_structure.py(102): ref[attribute_name] = attr --- modulename: namespace, funcname: __setitem__ namespace.py(58): self._attributes[k] = v metadata_container.py(378): with self._lock: metadata_container.py(370): attr.process_assignment(neptune_value, wait=wait) --- modulename: attribute, funcname: process_assignment attribute.py(63): if isinstance(value, ValueCopy): attribute.py(66): return self.assign(value, wait=wait) --- modulename: string, funcname: assign string.py(55): if not isinstance(value, StringVal): string.py(58): if len(value.value) > String.MAX_VALUE_LENGTH: string.py(72): with self._container.lock(): --- modulename: metadata_container, funcname: lock metadata_container.py(417): return self._lock string.py(73): self._enqueue_operation(self.create_assignment_operation(self._path, value.value), wait=wait) --- modulename: string, funcname: create_assignment_operation string.py(42): return AssignString(path, value) --- modulename: operation, funcname: __init__ (3): (4): --- modulename: attribute, funcname: _enqueue_operation attribute.py(45): self._container._op_processor.enqueue_operation(operation, wait=wait) --- modulename: async_operation_processor, funcname: enqueue_operation async_operation_processor.py(103): if self._drop_operations: async_operation_processor.py(105): self._last_version = self._queue.put(op) --- modulename: disk_queue, funcname: put disk_queue.py(92): version = self._last_put_file.read_local() + 1 --- modulename: sync_offset_file, funcname: read_local sync_offset_file.py(44): return self._last disk_queue.py(93): _json = json.dumps(self._serialize(obj, version)) --- modulename: disk_queue, funcname: _serialize disk_queue.py(246): return {"obj": self._to_dict(obj), "version": version} --- modulename: async_operation_processor, funcname: async_operation_processor.py(68): to_dict=lambda x: x.to_dict(), --- modulename: operation, funcname: to_dict operation.py(142): ret = super().to_dict() --- modulename: operation, funcname: to_dict operation.py(67): return {"type": self.__class__.__name__, "path": self.path} operation.py(143): ret["value"] = self.value operation.py(144): return ret --- modulename: __init__, funcname: dumps __init__.py(227): if (not skipkeys and ensure_ascii and __init__.py(228): check_circular and allow_nan and __init__.py(227): if (not skipkeys and ensure_ascii and __init__.py(228): check_circular and allow_nan and __init__.py(227): if (not skipkeys and ensure_ascii and __init__.py(229): cls is None and indent is None and separators is None and __init__.py(230): default is None and not sort_keys and not kw): __init__.py(231): return _default_encoder.encode(obj) --- modulename: encoder, funcname: encode encoder.py(191): if isinstance(o, str): encoder.py(199): chunks = self.iterencode(o, _one_shot=True) --- modulename: encoder, funcname: iterencode encoder.py(214): if self.check_circular: encoder.py(215): markers = {} encoder.py(218): if self.ensure_ascii: encoder.py(219): _encoder = encode_basestring_ascii encoder.py(223): def floatstr(o, allow_nan=self.allow_nan, encoder.py(224): _repr=float.__repr__, _inf=INFINITY, _neginf=-INFINITY): encoder.py(223): def floatstr(o, allow_nan=self.allow_nan, encoder.py(246): if (_one_shot and c_make_encoder is not None encoder.py(247): and self.indent is None): encoder.py(248): _iterencode = c_make_encoder( encoder.py(249): markers, self.default, _encoder, self.indent, encoder.py(250): self.key_separator, self.item_separator, self.sort_keys, encoder.py(251): self.skipkeys, self.allow_nan) encoder.py(248): _iterencode = c_make_encoder( encoder.py(257): return _iterencode(o, 0) encoder.py(200): if not isinstance(chunks, (list, tuple)): encoder.py(202): return ''.join(chunks) disk_queue.py(94): if self._file_size + len(_json) > self._max_file_size: disk_queue.py(100): self._writer.write(_json + "\n") disk_queue.py(101): self._last_put_file.write(version) --- modulename: sync_offset_file, funcname: write sync_offset_file.py(30): self._file.seek(0) --- modulename: codecs, funcname: reset codecs.py(328): IncrementalDecoder.reset(self) --- modulename: codecs, funcname: reset codecs.py(276): def reset(self): codecs.py(329): self.buffer = b"" --- modulename: codecs, funcname: reset codecs.py(203): def reset(self): sync_offset_file.py(31): self._file.write(str(offset)) --- modulename: codecs, funcname: reset codecs.py(328): IncrementalDecoder.reset(self) --- modulename: codecs, funcname: reset codecs.py(276): def reset(self): codecs.py(329): self.buffer = b"" sync_offset_file.py(32): self._file.truncate() sync_offset_file.py(33): self._file.flush() sync_offset_file.py(34): self._last = offset disk_queue.py(102): self._file_size += len(_json) + 1 disk_queue.py(103): return version async_operation_processor.py(106): if self._queue.size() > self._batch_size / 2: --- modulename: disk_queue, funcname: size disk_queue.py(223): return self._last_put_file.read_local() - self._last_ack_file.read_local() --- modulename: sync_offset_file, funcname: read_local sync_offset_file.py(44): return self._last --- modulename: sync_offset_file, funcname: read_local sync_offset_file.py(44): return self._last async_operation_processor.py(108): if wait: string.py(72): with self._container.lock(): metadata_container.py(371): return attr metadata_container.py(358): with self._lock: handler.py(208): with self._container.lock(): source_code.py(62): if source_files is not None: source_code.py(63): run[attr_consts.SOURCE_CODE_FILES_ATTRIBUTE_PATH].upload_files(source_files) --- modulename: metadata_container, funcname: inner_fun metadata_container.py(83): self._raise_if_stopped() --- modulename: run, funcname: _raise_if_stopped run.py(459): if self._state == ContainerState.STOPPED: metadata_container.py(84): return fun(self, *args, **kwargs) --- modulename: metadata_container, funcname: __getitem__ metadata_container.py(184): return Handler(self, path) --- modulename: handler, funcname: __init__ handler.py(101): super().__init__() handler.py(102): self._container = container handler.py(103): self._path = path --- modulename: handler, funcname: inner_fun handler.py(85): validate_path_not_protected(self._path, self) --- modulename: handler, funcname: validate_path_not_protected handler.py(77): path_protection_exception = handler._PROTECTED_PATHS.get(target_path) handler.py(78): if path_protection_exception: handler.py(86): return fun(self, *args, **kwargs) --- modulename: handler, funcname: upload_files handler.py(260): if is_collection(value): --- modulename: __init__, funcname: is_collection __init__.py(137): return isinstance(var, (list, set, tuple)) handler.py(263): verify_type("value", value, str) --- modulename: __init__, funcname: verify_type __init__.py(66): try: __init__.py(67): if isinstance(expected_type, tuple): __init__.py(70): type_name = get_type_name(expected_type) --- modulename: __init__, funcname: get_type_name __init__.py(127): return _type.__name__ if hasattr(_type, "__name__") else str(_type) __init__.py(75): if not isinstance(var, expected_type): __init__.py(78): if isinstance(var, IOBase) and not hasattr(var, "read"): --- modulename: abc, funcname: __instancecheck__ abc.py(119): return _abc_instancecheck(cls, instance) handler.py(265): with self._container.lock(): --- modulename: metadata_container, funcname: lock metadata_container.py(417): return self._lock handler.py(266): attr = self._container.get_attribute(self._path) --- modulename: metadata_container, funcname: get_attribute metadata_container.py(374): with self._lock: metadata_container.py(375): return self._structure.get(parse_path(path)) --- modulename: paths, funcname: parse_path paths.py(26): return _remove_empty_paths(path.split("/")) --- modulename: paths, funcname: _remove_empty_paths paths.py(22): return list(filter(bool, paths)) --- modulename: container_structure, funcname: get container_structure.py(68): ref = self._structure container_structure.py(70): for index, part in enumerate(path): container_structure.py(71): if not isinstance(ref, self._node_type): container_structure.py(76): if part not in ref: --- modulename: _collections_abc, funcname: __contains__ _collections_abc.py(829): try: _collections_abc.py(830): self[key] --- modulename: namespace, funcname: __getitem__ namespace.py(64): return self._attributes[k] _collections_abc.py(834): return True container_structure.py(78): ref = ref[part] --- modulename: namespace, funcname: __getitem__ namespace.py(64): return self._attributes[k] container_structure.py(70): for index, part in enumerate(path): container_structure.py(71): if not isinstance(ref, self._node_type): container_structure.py(76): if part not in ref: --- modulename: _collections_abc, funcname: __contains__ _collections_abc.py(829): try: _collections_abc.py(830): self[key] --- modulename: namespace, funcname: __getitem__ namespace.py(64): return self._attributes[k] _collections_abc.py(831): except KeyError: _collections_abc.py(832): return False container_structure.py(77): return None metadata_container.py(374): with self._lock: handler.py(267): if attr is None: handler.py(268): attr = FileSet(self._container, parse_path(self._path)) --- modulename: paths, funcname: parse_path paths.py(26): return _remove_empty_paths(path.split("/")) --- modulename: paths, funcname: _remove_empty_paths paths.py(22): return list(filter(bool, paths)) --- modulename: attribute, funcname: __init__ attribute.py(37): super().__init__() attribute.py(38): self._container = container attribute.py(39): self._path = path handler.py(269): self._container.set_attribute(self._path, attr) --- modulename: metadata_container, funcname: set_attribute metadata_container.py(378): with self._lock: metadata_container.py(379): return self._structure.set(parse_path(path), attribute) --- modulename: paths, funcname: parse_path paths.py(26): return _remove_empty_paths(path.split("/")) --- modulename: paths, funcname: _remove_empty_paths paths.py(22): return list(filter(bool, paths)) --- modulename: container_structure, funcname: set container_structure.py(83): ref = self._structure container_structure.py(84): location, attribute_name = path[:-1], path[-1] container_structure.py(86): for idx, part in enumerate(location): container_structure.py(87): if part not in ref: --- modulename: _collections_abc, funcname: __contains__ _collections_abc.py(829): try: _collections_abc.py(830): self[key] --- modulename: namespace, funcname: __getitem__ namespace.py(64): return self._attributes[k] _collections_abc.py(834): return True container_structure.py(89): ref = ref[part] --- modulename: namespace, funcname: __getitem__ namespace.py(64): return self._attributes[k] container_structure.py(90): if not isinstance(ref, self._node_type): container_structure.py(86): for idx, part in enumerate(location): container_structure.py(96): if attribute_name in ref and isinstance(ref[attribute_name], self._node_type): --- modulename: _collections_abc, funcname: __contains__ _collections_abc.py(829): try: _collections_abc.py(830): self[key] --- modulename: namespace, funcname: __getitem__ namespace.py(64): return self._attributes[k] _collections_abc.py(831): except KeyError: _collections_abc.py(832): return False container_structure.py(102): ref[attribute_name] = attr --- modulename: namespace, funcname: __setitem__ namespace.py(58): self._attributes[k] = v metadata_container.py(378): with self._lock: handler.py(270): attr.upload_files(value, wait=wait) --- modulename: file_set, funcname: upload_files file_set.py(49): if isinstance(globs, str): file_set.py(50): globs = [globs] file_set.py(53): self._enqueue_upload_operation(globs, reset=False, wait=wait) --- modulename: file_set, funcname: _enqueue_upload_operation file_set.py(64): with self._container.lock(): --- modulename: metadata_container, funcname: lock metadata_container.py(417): return self._lock file_set.py(65): abs_file_globs = list(os.path.abspath(file_glob) for file_glob in globs) --- modulename: file_set, funcname: file_set.py(65): abs_file_globs = list(os.path.abspath(file_glob) for file_glob in globs) --- modulename: posixpath, funcname: abspath posixpath.py(379): path = os.fspath(path) posixpath.py(380): if not isabs(path): --- modulename: posixpath, funcname: isabs posixpath.py(62): s = os.fspath(s) posixpath.py(63): sep = _get_sep(s) --- modulename: posixpath, funcname: _get_sep posixpath.py(42): if isinstance(path, bytes): posixpath.py(45): return '/' posixpath.py(64): return s.startswith(sep) posixpath.py(386): return normpath(path) --- modulename: posixpath, funcname: normpath posixpath.py(340): path = os.fspath(path) posixpath.py(341): if isinstance(path, bytes): posixpath.py(347): sep = '/' posixpath.py(348): empty = '' posixpath.py(349): dot = '.' posixpath.py(350): dotdot = '..' posixpath.py(351): if path == empty: posixpath.py(353): initial_slashes = path.startswith(sep) posixpath.py(357): if (initial_slashes and posixpath.py(358): path.startswith(sep*2) and not path.startswith(sep*3)): posixpath.py(357): if (initial_slashes and posixpath.py(360): comps = path.split(sep) posixpath.py(361): new_comps = [] posixpath.py(362): for comp in comps: posixpath.py(363): if comp in (empty, dot): posixpath.py(364): continue posixpath.py(362): for comp in comps: posixpath.py(363): if comp in (empty, dot): posixpath.py(365): if (comp != dotdot or (not initial_slashes and not new_comps) or posixpath.py(367): new_comps.append(comp) posixpath.py(362): for comp in comps: posixpath.py(363): if comp in (empty, dot): posixpath.py(365): if (comp != dotdot or (not initial_slashes and not new_comps) or posixpath.py(367): new_comps.append(comp) posixpath.py(362): for comp in comps: posixpath.py(363): if comp in (empty, dot): posixpath.py(365): if (comp != dotdot or (not initial_slashes and not new_comps) or posixpath.py(367): new_comps.append(comp) posixpath.py(362): for comp in comps: posixpath.py(363): if comp in (empty, dot): posixpath.py(365): if (comp != dotdot or (not initial_slashes and not new_comps) or posixpath.py(367): new_comps.append(comp) posixpath.py(362): for comp in comps: posixpath.py(363): if comp in (empty, dot): posixpath.py(365): if (comp != dotdot or (not initial_slashes and not new_comps) or posixpath.py(367): new_comps.append(comp) posixpath.py(362): for comp in comps: posixpath.py(363): if comp in (empty, dot): posixpath.py(365): if (comp != dotdot or (not initial_slashes and not new_comps) or posixpath.py(367): new_comps.append(comp) posixpath.py(362): for comp in comps: posixpath.py(363): if comp in (empty, dot): posixpath.py(365): if (comp != dotdot or (not initial_slashes and not new_comps) or posixpath.py(367): new_comps.append(comp) posixpath.py(362): for comp in comps: posixpath.py(363): if comp in (empty, dot): posixpath.py(365): if (comp != dotdot or (not initial_slashes and not new_comps) or posixpath.py(367): new_comps.append(comp) posixpath.py(362): for comp in comps: posixpath.py(363): if comp in (empty, dot): posixpath.py(365): if (comp != dotdot or (not initial_slashes and not new_comps) or posixpath.py(367): new_comps.append(comp) posixpath.py(362): for comp in comps: posixpath.py(370): comps = new_comps posixpath.py(371): path = sep.join(comps) posixpath.py(372): if initial_slashes: posixpath.py(373): path = sep*initial_slashes + path posixpath.py(374): return path or dot --- modulename: file_set, funcname: file_set.py(65): abs_file_globs = list(os.path.abspath(file_glob) for file_glob in globs) file_set.py(66): self._enqueue_operation(UploadFileSet(self._path, abs_file_globs, reset=reset), wait=wait) --- modulename: operation, funcname: __init__ (3): (4): (5): --- modulename: attribute, funcname: _enqueue_operation attribute.py(45): self._container._op_processor.enqueue_operation(operation, wait=wait) --- modulename: async_operation_processor, funcname: enqueue_operation async_operation_processor.py(103): if self._drop_operations: async_operation_processor.py(105): self._last_version = self._queue.put(op) --- modulename: disk_queue, funcname: put disk_queue.py(92): version = self._last_put_file.read_local() + 1 --- modulename: sync_offset_file, funcname: read_local sync_offset_file.py(44): return self._last disk_queue.py(93): _json = json.dumps(self._serialize(obj, version)) --- modulename: disk_queue, funcname: _serialize disk_queue.py(246): return {"obj": self._to_dict(obj), "version": version} --- modulename: async_operation_processor, funcname: async_operation_processor.py(68): to_dict=lambda x: x.to_dict(), --- modulename: operation, funcname: to_dict operation.py(284): ret = super().to_dict() --- modulename: operation, funcname: to_dict operation.py(67): return {"type": self.__class__.__name__, "path": self.path} operation.py(285): ret["file_globs"] = self.file_globs operation.py(286): ret["reset"] = str(self.reset) operation.py(287): return ret --- modulename: __init__, funcname: dumps __init__.py(227): if (not skipkeys and ensure_ascii and __init__.py(228): check_circular and allow_nan and __init__.py(227): if (not skipkeys and ensure_ascii and __init__.py(228): check_circular and allow_nan and __init__.py(227): if (not skipkeys and ensure_ascii and __init__.py(229): cls is None and indent is None and separators is None and __init__.py(230): default is None and not sort_keys and not kw): __init__.py(231): return _default_encoder.encode(obj) --- modulename: encoder, funcname: encode encoder.py(191): if isinstance(o, str): encoder.py(199): chunks = self.iterencode(o, _one_shot=True) --- modulename: encoder, funcname: iterencode encoder.py(214): if self.check_circular: encoder.py(215): markers = {} encoder.py(218): if self.ensure_ascii: encoder.py(219): _encoder = encode_basestring_ascii encoder.py(223): def floatstr(o, allow_nan=self.allow_nan, encoder.py(224): _repr=float.__repr__, _inf=INFINITY, _neginf=-INFINITY): encoder.py(223): def floatstr(o, allow_nan=self.allow_nan, encoder.py(246): if (_one_shot and c_make_encoder is not None encoder.py(247): and self.indent is None): encoder.py(248): _iterencode = c_make_encoder( encoder.py(249): markers, self.default, _encoder, self.indent, encoder.py(250): self.key_separator, self.item_separator, self.sort_keys, encoder.py(251): self.skipkeys, self.allow_nan) encoder.py(248): _iterencode = c_make_encoder( encoder.py(257): return _iterencode(o, 0) encoder.py(200): if not isinstance(chunks, (list, tuple)): encoder.py(202): return ''.join(chunks) disk_queue.py(94): if self._file_size + len(_json) > self._max_file_size: disk_queue.py(100): self._writer.write(_json + "\n") disk_queue.py(101): self._last_put_file.write(version) --- modulename: sync_offset_file, funcname: write sync_offset_file.py(30): self._file.seek(0) --- modulename: codecs, funcname: reset codecs.py(328): IncrementalDecoder.reset(self) --- modulename: codecs, funcname: reset codecs.py(276): def reset(self): codecs.py(329): self.buffer = b"" --- modulename: codecs, funcname: reset codecs.py(203): def reset(self): sync_offset_file.py(31): self._file.write(str(offset)) --- modulename: codecs, funcname: reset codecs.py(328): IncrementalDecoder.reset(self) --- modulename: codecs, funcname: reset codecs.py(276): def reset(self): codecs.py(329): self.buffer = b"" sync_offset_file.py(32): self._file.truncate() sync_offset_file.py(33): self._file.flush() sync_offset_file.py(34): self._last = offset disk_queue.py(102): self._file_size += len(_json) + 1 disk_queue.py(103): return version async_operation_processor.py(106): if self._queue.size() > self._batch_size / 2: --- modulename: disk_queue, funcname: size disk_queue.py(223): return self._last_put_file.read_local() - self._last_ack_file.read_local() --- modulename: sync_offset_file, funcname: read_local sync_offset_file.py(44): return self._last --- modulename: sync_offset_file, funcname: read_local sync_offset_file.py(44): return self._last async_operation_processor.py(108): if wait: file_set.py(64): with self._container.lock(): handler.py(265): with self._container.lock(): metadata_container.py(144): self._startup(debug_mode=mode == Mode.DEBUG) --- modulename: metadata_container, funcname: _startup metadata_container.py(483): if not debug_mode: metadata_container.py(484): logger.info(self.get_url()) --- modulename: run, funcname: get_url run.py(464): return self._backend.get_run_url( run.py(465): run_id=self._id, run.py(466): workspace=self._workspace, run.py(467): project_name=self._project_name, run.py(468): sys_id=self._sys_id, run.py(464): return self._backend.get_run_url( --- modulename: hosted_neptune_backend, funcname: get_run_url hosted_neptune_backend.py(1052): base_url = self.get_display_address() --- modulename: hosted_neptune_backend, funcname: get_display_address hosted_neptune_backend.py(174): return self._client_config.display_url hosted_neptune_backend.py(1053): return f"{base_url}/{workspace}/{project_name}/e/{sys_id}" --- modulename: __init__, funcname: info __init__.py(1476): if self.isEnabledFor(INFO): --- modulename: __init__, funcname: isEnabledFor __init__.py(1728): if self.disabled: __init__.py(1731): try: __init__.py(1732): return self._cache[level] __init__.py(1733): except KeyError: __init__.py(1734): _acquireLock() --- modulename: __init__, funcname: _acquireLock __init__.py(225): if _lock: __init__.py(226): _lock.acquire() __init__.py(1735): try: __init__.py(1736): if self.manager.disable >= level: --- modulename: __init__, funcname: disable __init__.py(1309): return self._disable __init__.py(1740): level >= self.getEffectiveLevel() --- modulename: __init__, funcname: getEffectiveLevel __init__.py(1717): logger = self __init__.py(1718): while logger: __init__.py(1719): if logger.level: __init__.py(1720): return logger.level __init__.py(1739): is_enabled = self._cache[level] = ( __init__.py(1743): _releaseLock() --- modulename: __init__, funcname: _releaseLock __init__.py(232): if _lock: __init__.py(233): _lock.release() __init__.py(1744): return is_enabled __init__.py(1477): self._log(INFO, msg, args, **kwargs) --- modulename: __init__, funcname: _log __init__.py(1606): sinfo = None __init__.py(1607): if _srcfile: __init__.py(1611): try: __init__.py(1612): fn, lno, func, sinfo = self.findCaller(stack_info, stacklevel) --- modulename: __init__, funcname: findCaller __init__.py(1554): f = currentframe() --- modulename: __init__, funcname: __init__.py(160): currentframe = lambda: sys._getframe(3) __init__.py(1557): if f is not None: __init__.py(1558): f = f.f_back __init__.py(1559): orig_f = f __init__.py(1560): while f and stacklevel > 1: __init__.py(1563): if not f: __init__.py(1565): rv = "(unknown file)", 0, "(unknown function)", None __init__.py(1566): while hasattr(f, "f_code"): __init__.py(1567): co = f.f_code __init__.py(1568): filename = os.path.normcase(co.co_filename) --- modulename: posixpath, funcname: normcase posixpath.py(54): return os.fspath(s) __init__.py(1569): if filename == _srcfile: __init__.py(1572): sinfo = None __init__.py(1573): if stack_info: __init__.py(1581): rv = (co.co_filename, f.f_lineno, co.co_name, sinfo) __init__.py(1582): break __init__.py(1583): return rv __init__.py(1617): if exc_info: __init__.py(1622): record = self.makeRecord(self.name, level, fn, lno, msg, args, __init__.py(1623): exc_info, func, extra, sinfo) __init__.py(1622): record = self.makeRecord(self.name, level, fn, lno, msg, args, --- modulename: __init__, funcname: makeRecord __init__.py(1591): rv = _logRecordFactory(name, level, fn, lno, msg, args, exc_info, func, __init__.py(1592): sinfo) __init__.py(1591): rv = _logRecordFactory(name, level, fn, lno, msg, args, exc_info, func, --- modulename: __init__, funcname: __init__ __init__.py(288): ct = time.time() __init__.py(289): self.name = name __init__.py(290): self.msg = msg __init__.py(309): if (args and len(args) == 1 and isinstance(args[0], collections.abc.Mapping) __init__.py(312): self.args = args __init__.py(313): self.levelname = getLevelName(level) --- modulename: __init__, funcname: getLevelName __init__.py(138): result = _levelToName.get(level) __init__.py(139): if result is not None: __init__.py(140): return result __init__.py(314): self.levelno = level __init__.py(315): self.pathname = pathname __init__.py(316): try: __init__.py(317): self.filename = os.path.basename(pathname) --- modulename: posixpath, funcname: basename posixpath.py(142): p = os.fspath(p) posixpath.py(143): sep = _get_sep(p) --- modulename: posixpath, funcname: _get_sep posixpath.py(42): if isinstance(path, bytes): posixpath.py(45): return '/' posixpath.py(144): i = p.rfind(sep) + 1 posixpath.py(145): return p[i:] __init__.py(318): self.module = os.path.splitext(self.filename)[0] --- modulename: posixpath, funcname: splitext posixpath.py(118): p = os.fspath(p) posixpath.py(119): if isinstance(p, bytes): posixpath.py(123): sep = '/' posixpath.py(124): extsep = '.' posixpath.py(125): return genericpath._splitext(p, sep, None, extsep) --- modulename: genericpath, funcname: _splitext genericpath.py(128): sepIndex = p.rfind(sep) genericpath.py(129): if altsep: genericpath.py(133): dotIndex = p.rfind(extsep) genericpath.py(134): if dotIndex > sepIndex: genericpath.py(136): filenameIndex = sepIndex + 1 genericpath.py(137): while filenameIndex < dotIndex: genericpath.py(138): if p[filenameIndex:filenameIndex+1] != extsep: genericpath.py(139): return p[:dotIndex], p[dotIndex:] __init__.py(322): self.exc_info = exc_info __init__.py(323): self.exc_text = None # used to cache the traceback text __init__.py(324): self.stack_info = sinfo __init__.py(325): self.lineno = lineno __init__.py(326): self.funcName = func __init__.py(327): self.created = ct __init__.py(328): self.msecs = int((ct - int(ct)) * 1000) + 0.0 # see gh-89047 __init__.py(329): self.relativeCreated = (self.created - _startTime) * 1000 __init__.py(330): if logThreads: __init__.py(331): self.thread = threading.get_ident() __init__.py(332): self.threadName = threading.current_thread().name --- modulename: threading, funcname: current_thread threading.py(1437): try: threading.py(1438): return _active[get_ident()] --- modulename: threading, funcname: name threading.py(1137): assert self._initialized, "Thread.__init__() not called" threading.py(1138): return self._name __init__.py(336): if not logMultiprocessing: # pragma: no cover __init__.py(339): self.processName = 'MainProcess' __init__.py(340): mp = sys.modules.get('multiprocessing') __init__.py(341): if mp is not None: __init__.py(346): try: __init__.py(347): self.processName = mp.current_process().name --- modulename: process, funcname: current_process process.py(41): return _current_process --- modulename: process, funcname: name process.py(191): return self._name __init__.py(350): if logProcesses and hasattr(os, 'getpid'): __init__.py(351): self.process = os.getpid() __init__.py(1593): if extra is not None: __init__.py(1598): return rv __init__.py(1624): self.handle(record) --- modulename: __init__, funcname: handle __init__.py(1633): if (not self.disabled) and self.filter(record): --- modulename: __init__, funcname: filter __init__.py(818): rv = True __init__.py(819): for f in self.filters: __init__.py(827): return rv __init__.py(1634): self.callHandlers(record) --- modulename: __init__, funcname: callHandlers __init__.py(1690): c = self __init__.py(1691): found = 0 __init__.py(1692): while c: __init__.py(1693): for hdlr in c.handlers: __init__.py(1694): found = found + 1 __init__.py(1695): if record.levelno >= hdlr.level: __init__.py(1696): hdlr.handle(record) --- modulename: __init__, funcname: handle __init__.py(964): rv = self.filter(record) --- modulename: __init__, funcname: filter __init__.py(818): rv = True __init__.py(819): for f in self.filters: __init__.py(827): return rv __init__.py(965): if rv: __init__.py(966): self.acquire() --- modulename: __init__, funcname: acquire __init__.py(916): if self.lock: __init__.py(917): self.lock.acquire() __init__.py(967): try: __init__.py(968): self.emit(record) --- modulename: __init__, funcname: emit __init__.py(1099): try: __init__.py(1100): msg = self.format(record) --- modulename: __init__, funcname: format __init__.py(939): if self.formatter: __init__.py(940): fmt = self.formatter __init__.py(943): return fmt.format(record) --- modulename: __init__, funcname: format __init__.py(678): record.message = record.getMessage() --- modulename: __init__, funcname: getMessage __init__.py(366): msg = str(self.msg) __init__.py(367): if self.args: __init__.py(369): return msg __init__.py(679): if self.usesTime(): --- modulename: __init__, funcname: usesTime __init__.py(647): return self._style.usesTime() --- modulename: __init__, funcname: usesTime __init__.py(424): return self._fmt.find(self.asctime_search) >= 0 __init__.py(681): s = self.formatMessage(record) --- modulename: __init__, funcname: formatMessage __init__.py(650): return self._style.format(record) --- modulename: __init__, funcname: format __init__.py(439): try: __init__.py(440): return self._format(record) --- modulename: __init__, funcname: _format __init__.py(432): if defaults := self._defaults: __init__.py(435): values = record.__dict__ __init__.py(436): return self._fmt % values __init__.py(682): if record.exc_info: __init__.py(687): if record.exc_text: __init__.py(691): if record.stack_info: __init__.py(695): return s __init__.py(1101): stream = self.stream --- modulename: logger, funcname: stream logger.py(39): return sys.stdout __init__.py(1103): stream.write(msg + self.terminator) https://ui.neptune.ai/neuro/Gaddy/e/GAD-333 __init__.py(1104): self.flush() --- modulename: __init__, funcname: flush __init__.py(1081): self.acquire() --- modulename: __init__, funcname: acquire __init__.py(916): if self.lock: __init__.py(917): self.lock.acquire() __init__.py(1082): try: __init__.py(1083): if self.stream and hasattr(self.stream, "flush"): --- modulename: logger, funcname: stream logger.py(39): return sys.stdout --- modulename: logger, funcname: stream logger.py(39): return sys.stdout __init__.py(1084): self.stream.flush() --- modulename: logger, funcname: stream logger.py(39): return sys.stdout __init__.py(1086): self.release() --- modulename: __init__, funcname: release __init__.py(923): if self.lock: __init__.py(924): self.lock.release() __init__.py(970): self.release() --- modulename: __init__, funcname: release __init__.py(923): if self.lock: __init__.py(924): self.lock.release() __init__.py(971): return rv __init__.py(1693): for hdlr in c.handlers: __init__.py(1697): if not c.propagate: __init__.py(1698): c = None #break out __init__.py(1692): while c: __init__.py(1701): if (found == 0): metadata_container.py(486): self.start() --- modulename: metadata_container, funcname: start metadata_container.py(259): atexit.register(self._shutdown_hook) metadata_container.py(260): self._op_processor.start() --- modulename: async_operation_processor, funcname: start async_operation_processor.py(128): self._consumer.start() --- modulename: threading, funcname: start threading.py(926): if not self._initialized: threading.py(929): if self._started.is_set(): --- modulename: threading, funcname: is_set threading.py(555): return self._flag threading.py(932): with _active_limbo_lock: threading.py(933): _limbo[self] = self threading.py(932): with _active_limbo_lock: threading.py(934): try: threading.py(935): _start_new_thread(self._bootstrap, ()) threading.py(940): self._started.wait() --- modulename: threading, funcname: wait threading.py(604): with self._cond: --- modulename: threading, funcname: __enter__ threading.py(265): return self._lock.__enter__() threading.py(605): signaled = self._flag threading.py(606): if not signaled: threading.py(607): signaled = self._cond.wait(timeout) --- modulename: threading, funcname: wait threading.py(311): if not self._is_owned(): --- modulename: threading, funcname: _is_owned threading.py(282): if self._lock.acquire(False): threading.py(286): return True threading.py(313): waiter = _allocate_lock() threading.py(314): waiter.acquire() threading.py(315): self._waiters.append(waiter) threading.py(316): saved_state = self._release_save() --- modulename: threading, funcname: _release_save threading.py(274): self._lock.release() # No state to save threading.py(317): gotit = False threading.py(318): try: # restore state no matter what (e.g., KeyboardInterrupt) threading.py(319): if timeout is None: threading.py(320): waiter.acquire() --- modulename: async_operation_processor, funcname: run threading.py(321): gotit = True async_operation_processor.py(229): try: threading.py(327): return gotit async_operation_processor.py(230): super().run() threading.py(329): self._acquire_restore(saved_state) --- modulename: daemon, funcname: run --- modulename: threading, funcname: _acquire_restore daemon.py(50): self._is_running = True threading.py(277): self._lock.acquire() # Ignore saved state daemon.py(51): try: threading.py(330): if not gotit: daemon.py(52): while not self._interrupted: threading.py(608): return signaled daemon.py(53): self.work() threading.py(604): with self._cond: --- modulename: async_operation_processor, funcname: work --- modulename: threading, funcname: __exit__ async_operation_processor.py(237): ts = time() threading.py(268): return self._lock.__exit__(*args) async_operation_processor.py(238): if ts - self._last_flush >= self._sleep_time: metadata_container.py(261): self._bg_job.start(self) async_operation_processor.py(239): self._last_flush = ts --- modulename: backgroud_job_list, funcname: start async_operation_processor.py(240): self._processor._queue.flush() backgroud_job_list.py(36): for job in self._jobs: --- modulename: disk_queue, funcname: flush backgroud_job_list.py(37): job.start(container) disk_queue.py(165): self._writer.flush() --- modulename: ping_background_job, funcname: start ping_background_job.py(40): self._thread = self.ReportingThread(self._period, container) --- modulename: ping_background_job, funcname: __init__ ping_background_job.py(56): super().__init__(sleep_time=period, name="NeptunePing") --- modulename: daemon, funcname: __init__ daemon.py(29): super().__init__(daemon=True, name=name) --- modulename: threading, funcname: __init__ threading.py(850): assert group is None, "group argument must be None for now" threading.py(851): if kwargs is None: threading.py(852): kwargs = {} threading.py(853): if name: threading.py(854): name = str(name) threading.py(864): self._target = target threading.py(865): self._name = name threading.py(866): self._args = args threading.py(867): self._kwargs = kwargs threading.py(868): if daemon is not None: threading.py(869): self._daemonic = daemon threading.py(872): self._ident = None threading.py(873): if _HAVE_THREAD_NATIVE_ID: threading.py(874): self._native_id = None threading.py(875): self._tstate_lock = None threading.py(876): self._started = Event() --- modulename: threading, funcname: __init__ threading.py(546): self._cond = Condition(Lock()) disk_queue.py(166): self._last_ack_file.flush() --- modulename: threading, funcname: __init__ --- modulename: sync_offset_file, funcname: flush threading.py(237): if lock is None: sync_offset_file.py(47): self._file.flush() threading.py(239): self._lock = lock disk_queue.py(167): self._last_put_file.flush() threading.py(241): self.acquire = lock.acquire --- modulename: sync_offset_file, funcname: flush threading.py(242): self.release = lock.release sync_offset_file.py(47): self._file.flush() threading.py(246): try: threading.py(247): self._release_save = lock._release_save threading.py(248): except AttributeError: threading.py(249): pass threading.py(250): try: threading.py(251): self._acquire_restore = lock._acquire_restore async_operation_processor.py(242): while True: threading.py(252): except AttributeError: async_operation_processor.py(243): batch = self._processor._queue.get_batch(self._batch_size) threading.py(253): pass --- modulename: disk_queue, funcname: get_batch threading.py(254): try: disk_queue.py(144): if self._should_skip_to_ack: threading.py(255): self._is_owned = lock._is_owned disk_queue.py(145): first = self._skip_and_get() threading.py(256): except AttributeError: --- modulename: disk_queue, funcname: _skip_and_get threading.py(257): pass disk_queue.py(112): ack_version = self._last_ack_file.read_local() threading.py(258): self._waiters = _deque() --- modulename: sync_offset_file, funcname: read_local threading.py(547): self._flag = False sync_offset_file.py(44): return self._last threading.py(877): self._is_stopped = False disk_queue.py(113): while True: threading.py(878): self._initialized = True disk_queue.py(114): top_element = self._get() threading.py(880): self._stderr = _sys.stderr --- modulename: disk_queue, funcname: _get threading.py(881): self._invoke_excepthook = _make_invoke_excepthook() disk_queue.py(128): _json, size = self._reader.get_with_size() --- modulename: threading, funcname: _make_invoke_excepthook --- modulename: json_file_splitter, funcname: get_with_size threading.py(1306): old_excepthook = excepthook json_file_splitter.py(47): if self._parsed_queue: threading.py(1307): old_sys_excepthook = _sys.excepthook json_file_splitter.py(49): self._read_data() threading.py(1308): if old_excepthook is None: --- modulename: json_file_splitter, funcname: _read_data threading.py(1310): if old_sys_excepthook is None: json_file_splitter.py(55): if self._part_buffer.tell() < self.MAX_PART_READ: threading.py(1313): sys_exc_info = _sys.exc_info json_file_splitter.py(56): data = self._file.read(self.BUFFER_SIZE) threading.py(1314): local_print = print --- modulename: codecs, funcname: getstate threading.py(1315): local_sys = _sys codecs.py(333): return (self.buffer, 0) threading.py(1317): def invoke_excepthook(thread): threading.py(1349): return invoke_excepthook threading.py(883): _dangling.add(self) --- modulename: _weakrefset, funcname: add _weakrefset.py(87): if self._pending_removals: --- modulename: codecs, funcname: decode _weakrefset.py(89): self.data.add(ref(item, self._remove)) codecs.py(321): data = self.buffer + input daemon.py(30): self._sleep_time = sleep_time codecs.py(322): (result, consumed) = self._buffer_decode(data, self.errors, final) daemon.py(31): self._interrupted = False codecs.py(324): self.buffer = data[consumed:] daemon.py(32): self._event = threading.Event() codecs.py(325): return result --- modulename: threading, funcname: __init__ --- modulename: codecs, funcname: getstate threading.py(546): self._cond = Condition(Lock()) codecs.py(333): return (self.buffer, 0) --- modulename: threading, funcname: __init__ --- modulename: codecs, funcname: decode threading.py(237): if lock is None: codecs.py(321): data = self.buffer + input threading.py(239): self._lock = lock codecs.py(322): (result, consumed) = self._buffer_decode(data, self.errors, final) threading.py(241): self.acquire = lock.acquire codecs.py(324): self.buffer = data[consumed:] threading.py(242): self.release = lock.release codecs.py(325): return result threading.py(246): try: json_file_splitter.py(57): if not data: threading.py(247): self._release_save = lock._release_save json_file_splitter.py(59): if self._part_buffer.tell() > 0: threading.py(248): except AttributeError: json_file_splitter.py(61): self._decode(data) threading.py(249): pass --- modulename: json_file_splitter, funcname: _decode threading.py(250): try: json_file_splitter.py(72): start = self._json_start(data) threading.py(251): self._acquire_restore = lock._acquire_restore --- modulename: json_file_splitter, funcname: _json_start threading.py(252): except AttributeError: json_file_splitter.py(87): try: threading.py(253): pass json_file_splitter.py(88): return data.index("{", start) threading.py(254): try: json_file_splitter.py(73): while start is not None: threading.py(255): self._is_owned = lock._is_owned json_file_splitter.py(74): try: threading.py(256): except AttributeError: json_file_splitter.py(75): json_data, new_start = self._decoder.raw_decode(data, start) threading.py(257): pass --- modulename: decoder, funcname: raw_decode threading.py(258): self._waiters = _deque() decoder.py(352): try: threading.py(547): self._flag = False decoder.py(353): obj, end = self.scan_once(s, idx) daemon.py(33): self._is_running = False decoder.py(356): return obj, end daemon.py(34): self.last_backoff_time = 0 # used only with ConnectionRetryWrapper decorator json_file_splitter.py(76): size = new_start - start ping_background_job.py(57): self._container = container json_file_splitter.py(77): start = new_start ping_background_job.py(41): self._thread.start() json_file_splitter.py(82): self._parsed_queue.append((json_data, size)) --- modulename: threading, funcname: start json_file_splitter.py(83): start = self._json_start(data, start) threading.py(926): if not self._initialized: --- modulename: json_file_splitter, funcname: _json_start threading.py(929): if self._started.is_set(): json_file_splitter.py(87): try: --- modulename: threading, funcname: is_set json_file_splitter.py(88): return data.index("{", start) threading.py(555): return self._flag json_file_splitter.py(73): while start is not None: threading.py(932): with _active_limbo_lock: json_file_splitter.py(74): try: threading.py(933): _limbo[self] = self json_file_splitter.py(75): json_data, new_start = self._decoder.raw_decode(data, start) threading.py(932): with _active_limbo_lock: --- modulename: decoder, funcname: raw_decode threading.py(934): try: decoder.py(352): try: threading.py(935): _start_new_thread(self._bootstrap, ()) decoder.py(353): obj, end = self.scan_once(s, idx) threading.py(940): self._started.wait() decoder.py(356): return obj, end --- modulename: daemon, funcname: run json_file_splitter.py(76): size = new_start - start --- modulename: threading, funcname: wait json_file_splitter.py(77): start = new_start threading.py(604): with self._cond: json_file_splitter.py(82): self._parsed_queue.append((json_data, size)) --- modulename: threading, funcname: __enter__ json_file_splitter.py(83): start = self._json_start(data, start) threading.py(265): return self._lock.__enter__() --- modulename: json_file_splitter, funcname: _json_start threading.py(605): signaled = self._flag json_file_splitter.py(87): try: threading.py(606): if not signaled: json_file_splitter.py(88): return data.index("{", start) threading.py(608): return signaled json_file_splitter.py(73): while start is not None: threading.py(604): with self._cond: json_file_splitter.py(74): try: --- modulename: threading, funcname: __exit__ json_file_splitter.py(75): json_data, new_start = self._decoder.raw_decode(data, start) threading.py(268): return self._lock.__exit__(*args) --- modulename: decoder, funcname: raw_decode ping_background_job.py(42): self._started = True decoder.py(352): try: backgroud_job_list.py(36): for job in self._jobs: decoder.py(353): obj, end = self.scan_once(s, idx) backgroud_job_list.py(37): job.start(container) decoder.py(356): return obj, end --- modulename: websocket_signals_background_job, funcname: start json_file_splitter.py(76): size = new_start - start websocket_signals_background_job.py(54): self._thread = self._ListenerThread(container, self._ws_factory.create()) json_file_splitter.py(77): start = new_start --- modulename: websockets_factory, funcname: create json_file_splitter.py(82): self._parsed_queue.append((json_data, size)) websockets_factory.py(33): return ReconnectingWebsocket( json_file_splitter.py(83): start = self._json_start(data, start) websockets_factory.py(34): url=self._url, --- modulename: json_file_splitter, funcname: _json_start websockets_factory.py(35): oauth2_session=self._session, json_file_splitter.py(87): try: websockets_factory.py(36): shutdown_event=threading.Event(), json_file_splitter.py(88): return data.index("{", start) --- modulename: threading, funcname: __init__ json_file_splitter.py(73): while start is not None: threading.py(546): self._cond = Condition(Lock()) json_file_splitter.py(74): try: --- modulename: threading, funcname: __init__ json_file_splitter.py(75): json_data, new_start = self._decoder.raw_decode(data, start) threading.py(237): if lock is None: --- modulename: decoder, funcname: raw_decode threading.py(239): self._lock = lock decoder.py(352): try: threading.py(241): self.acquire = lock.acquire decoder.py(353): obj, end = self.scan_once(s, idx) threading.py(242): self.release = lock.release decoder.py(356): return obj, end threading.py(246): try: json_file_splitter.py(76): size = new_start - start threading.py(247): self._release_save = lock._release_save json_file_splitter.py(77): start = new_start threading.py(248): except AttributeError: json_file_splitter.py(82): self._parsed_queue.append((json_data, size)) threading.py(249): pass json_file_splitter.py(83): start = self._json_start(data, start) threading.py(250): try: --- modulename: json_file_splitter, funcname: _json_start threading.py(251): self._acquire_restore = lock._acquire_restore json_file_splitter.py(87): try: threading.py(252): except AttributeError: json_file_splitter.py(88): return data.index("{", start) threading.py(253): pass json_file_splitter.py(73): while start is not None: threading.py(254): try: json_file_splitter.py(74): try: threading.py(255): self._is_owned = lock._is_owned json_file_splitter.py(75): json_data, new_start = self._decoder.raw_decode(data, start) threading.py(256): except AttributeError: --- modulename: decoder, funcname: raw_decode threading.py(257): pass decoder.py(352): try: threading.py(258): self._waiters = _deque() decoder.py(353): obj, end = self.scan_once(s, idx) threading.py(547): self._flag = False decoder.py(356): return obj, end websockets_factory.py(37): proxies=self._proxies, json_file_splitter.py(76): size = new_start - start websockets_factory.py(33): return ReconnectingWebsocket( json_file_splitter.py(77): start = new_start --- modulename: reconnecting_websocket, funcname: __init__ json_file_splitter.py(82): self._parsed_queue.append((json_data, size)) reconnecting_websocket.py(32): self.url = url json_file_splitter.py(83): start = self._json_start(data, start) reconnecting_websocket.py(33): self.client = WebsocketClientAdapter() --- modulename: json_file_splitter, funcname: _json_start --- modulename: websocket_client_adapter, funcname: __init__ json_file_splitter.py(87): try: websocket_client_adapter.py(30): self._ws_client = None json_file_splitter.py(88): return data.index("{", start) reconnecting_websocket.py(34): self._shutdown_event = shutdown_event json_file_splitter.py(73): while start is not None: reconnecting_websocket.py(35): self._oauth2_session = oauth2_session json_file_splitter.py(74): try: reconnecting_websocket.py(36): self._reconnect_counter = ReconnectCounter() json_file_splitter.py(75): json_data, new_start = self._decoder.raw_decode(data, start) --- modulename: reconnecting_websocket, funcname: __init__ --- modulename: decoder, funcname: raw_decode reconnecting_websocket.py(95): self.retries = 0 decoder.py(352): try: reconnecting_websocket.py(37): self._token = oauth2_session.token decoder.py(353): obj, end = self.scan_once(s, idx) --- modulename: oauth2_session, funcname: token decoder.py(356): return obj, end oauth2_session.py(124): return getattr(self._client, "token", None) json_file_splitter.py(76): size = new_start - start reconnecting_websocket.py(38): self._proxies = proxies json_file_splitter.py(77): start = new_start --- modulename: websocket_signals_background_job, funcname: __init__ json_file_splitter.py(82): self._parsed_queue.append((json_data, size)) websocket_signals_background_job.py(73): super().__init__(sleep_time=0, name="NeptuneWebhooks") json_file_splitter.py(83): start = self._json_start(data, start) --- modulename: daemon, funcname: __init__ --- modulename: json_file_splitter, funcname: _json_start daemon.py(29): super().__init__(daemon=True, name=name) json_file_splitter.py(87): try: --- modulename: threading, funcname: __init__ json_file_splitter.py(88): return data.index("{", start) threading.py(850): assert group is None, "group argument must be None for now" json_file_splitter.py(73): while start is not None: threading.py(851): if kwargs is None: json_file_splitter.py(74): try: threading.py(852): kwargs = {} json_file_splitter.py(75): json_data, new_start = self._decoder.raw_decode(data, start) threading.py(853): if name: --- modulename: decoder, funcname: raw_decode threading.py(854): name = str(name) decoder.py(352): try: threading.py(864): self._target = target decoder.py(353): obj, end = self.scan_once(s, idx) threading.py(865): self._name = name decoder.py(356): return obj, end threading.py(866): self._args = args json_file_splitter.py(76): size = new_start - start threading.py(867): self._kwargs = kwargs json_file_splitter.py(77): start = new_start threading.py(868): if daemon is not None: json_file_splitter.py(82): self._parsed_queue.append((json_data, size)) threading.py(869): self._daemonic = daemon json_file_splitter.py(83): start = self._json_start(data, start) threading.py(872): self._ident = None --- modulename: json_file_splitter, funcname: _json_start threading.py(873): if _HAVE_THREAD_NATIVE_ID: json_file_splitter.py(87): try: threading.py(874): self._native_id = None json_file_splitter.py(88): return data.index("{", start) threading.py(875): self._tstate_lock = None json_file_splitter.py(73): while start is not None: threading.py(876): self._started = Event() json_file_splitter.py(74): try: --- modulename: threading, funcname: __init__ json_file_splitter.py(75): json_data, new_start = self._decoder.raw_decode(data, start) threading.py(546): self._cond = Condition(Lock()) --- modulename: decoder, funcname: raw_decode --- modulename: threading, funcname: __init__ decoder.py(352): try: threading.py(237): if lock is None: decoder.py(353): obj, end = self.scan_once(s, idx) threading.py(239): self._lock = lock decoder.py(356): return obj, end threading.py(241): self.acquire = lock.acquire json_file_splitter.py(76): size = new_start - start threading.py(242): self.release = lock.release json_file_splitter.py(77): start = new_start threading.py(246): try: json_file_splitter.py(82): self._parsed_queue.append((json_data, size)) threading.py(247): self._release_save = lock._release_save json_file_splitter.py(83): start = self._json_start(data, start) threading.py(248): except AttributeError: --- modulename: json_file_splitter, funcname: _json_start threading.py(249): pass json_file_splitter.py(87): try: threading.py(250): try: json_file_splitter.py(88): return data.index("{", start) threading.py(251): self._acquire_restore = lock._acquire_restore json_file_splitter.py(73): while start is not None: threading.py(252): except AttributeError: json_file_splitter.py(74): try: threading.py(253): pass json_file_splitter.py(75): json_data, new_start = self._decoder.raw_decode(data, start) threading.py(254): try: --- modulename: decoder, funcname: raw_decode threading.py(255): self._is_owned = lock._is_owned decoder.py(352): try: threading.py(256): except AttributeError: decoder.py(353): obj, end = self.scan_once(s, idx) threading.py(257): pass decoder.py(356): return obj, end threading.py(258): self._waiters = _deque() json_file_splitter.py(76): size = new_start - start threading.py(547): self._flag = False json_file_splitter.py(77): start = new_start threading.py(877): self._is_stopped = False json_file_splitter.py(82): self._parsed_queue.append((json_data, size)) threading.py(878): self._initialized = True daemon.py(50): self._is_running = True threading.py(880): self._stderr = _sys.stderr threading.py(881): self._invoke_excepthook = _make_invoke_excepthook() json_file_splitter.py(83): start = self._json_start(data, start) daemon.py(51): try: --- modulename: json_file_splitter, funcname: _json_start --- modulename: threading, funcname: _make_invoke_excepthook json_file_splitter.py(87): try: threading.py(1306): old_excepthook = excepthook json_file_splitter.py(88): return data.index("{", start) threading.py(1307): old_sys_excepthook = _sys.excepthook json_file_splitter.py(73): while start is not None: threading.py(1308): if old_excepthook is None: json_file_splitter.py(74): try: threading.py(1310): if old_sys_excepthook is None: json_file_splitter.py(75): json_data, new_start = self._decoder.raw_decode(data, start) threading.py(1313): sys_exc_info = _sys.exc_info --- modulename: decoder, funcname: raw_decode threading.py(1314): local_print = print decoder.py(352): try: threading.py(1315): local_sys = _sys decoder.py(353): obj, end = self.scan_once(s, idx) threading.py(1317): def invoke_excepthook(thread): decoder.py(356): return obj, end threading.py(1349): return invoke_excepthook json_file_splitter.py(76): size = new_start - start threading.py(883): _dangling.add(self) json_file_splitter.py(77): start = new_start --- modulename: _weakrefset, funcname: add json_file_splitter.py(82): self._parsed_queue.append((json_data, size)) _weakrefset.py(87): if self._pending_removals: json_file_splitter.py(83): start = self._json_start(data, start) _weakrefset.py(89): self.data.add(ref(item, self._remove)) --- modulename: json_file_splitter, funcname: _json_start daemon.py(30): self._sleep_time = sleep_time json_file_splitter.py(87): try: daemon.py(31): self._interrupted = False json_file_splitter.py(88): return data.index("{", start) daemon.py(32): self._event = threading.Event() json_file_splitter.py(73): while start is not None: --- modulename: threading, funcname: __init__ json_file_splitter.py(74): try: threading.py(546): self._cond = Condition(Lock()) json_file_splitter.py(75): json_data, new_start = self._decoder.raw_decode(data, start) --- modulename: threading, funcname: __init__ --- modulename: decoder, funcname: raw_decode threading.py(237): if lock is None: decoder.py(352): try: threading.py(239): self._lock = lock decoder.py(353): obj, end = self.scan_once(s, idx) threading.py(241): self.acquire = lock.acquire decoder.py(356): return obj, end threading.py(242): self.release = lock.release json_file_splitter.py(76): size = new_start - start threading.py(246): try: json_file_splitter.py(77): start = new_start threading.py(247): self._release_save = lock._release_save json_file_splitter.py(82): self._parsed_queue.append((json_data, size)) threading.py(248): except AttributeError: json_file_splitter.py(83): start = self._json_start(data, start) threading.py(249): pass --- modulename: json_file_splitter, funcname: _json_start threading.py(250): try: json_file_splitter.py(87): try: threading.py(251): self._acquire_restore = lock._acquire_restore json_file_splitter.py(88): return data.index("{", start) threading.py(252): except AttributeError: json_file_splitter.py(73): while start is not None: threading.py(253): pass json_file_splitter.py(74): try: threading.py(254): try: json_file_splitter.py(75): json_data, new_start = self._decoder.raw_decode(data, start) threading.py(255): self._is_owned = lock._is_owned --- modulename: decoder, funcname: raw_decode threading.py(256): except AttributeError: decoder.py(352): try: threading.py(257): pass decoder.py(353): obj, end = self.scan_once(s, idx) threading.py(258): self._waiters = _deque() decoder.py(356): return obj, end threading.py(547): self._flag = False json_file_splitter.py(76): size = new_start - start daemon.py(33): self._is_running = False json_file_splitter.py(77): start = new_start daemon.py(34): self.last_backoff_time = 0 # used only with ConnectionRetryWrapper decorator json_file_splitter.py(82): self._parsed_queue.append((json_data, size)) websocket_signals_background_job.py(74): self._container = container json_file_splitter.py(83): start = self._json_start(data, start) websocket_signals_background_job.py(75): self._ws_client = ws_client --- modulename: json_file_splitter, funcname: _json_start websocket_signals_background_job.py(55): self._thread.start() json_file_splitter.py(87): try: --- modulename: threading, funcname: start json_file_splitter.py(88): return data.index("{", start) threading.py(926): if not self._initialized: json_file_splitter.py(89): except ValueError: threading.py(929): if self._started.is_set(): json_file_splitter.py(90): return None --- modulename: threading, funcname: is_set json_file_splitter.py(73): while start is not None: threading.py(555): return self._flag json_file_splitter.py(63): if not self._parsed_queue: threading.py(932): with _active_limbo_lock: json_file_splitter.py(50): if self._parsed_queue: threading.py(933): _limbo[self] = self json_file_splitter.py(51): return self._parsed_queue.popleft() threading.py(932): with _active_limbo_lock: disk_queue.py(129): if not _json: threading.py(934): try: disk_queue.py(137): try: threading.py(935): _start_new_thread(self._bootstrap, ()) disk_queue.py(138): obj, ver = self._deserialize(_json) threading.py(940): self._started.wait() --- modulename: disk_queue, funcname: _deserialize --- modulename: threading, funcname: wait disk_queue.py(249): return self._from_dict(data["obj"]), data["version"] --- modulename: daemon, funcname: run daemon.py(52): while not self._interrupted: threading.py(604): with self._cond: daemon.py(50): self._is_running = True --- modulename: operation, funcname: from_dict daemon.py(53): self.work() daemon.py(51): try: --- modulename: threading, funcname: __enter__ --- modulename: daemon, funcname: wrapper daemon.py(52): while not self._interrupted: operation.py(71): if "type" not in data: threading.py(265): return self._lock.__enter__() daemon.py(74): while not self_._interrupted: daemon.py(53): self.work() operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} threading.py(605): signaled = self._flag daemon.py(75): try: --- modulename: websocket_signals_background_job, funcname: work --- modulename: operation, funcname: all_subclasses threading.py(606): if not signaled: daemon.py(76): result = func(self_, *args, **kwargs) websocket_signals_background_job.py(78): try: operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(608): return signaled --- modulename: ping_background_job, funcname: work websocket_signals_background_job.py(79): raw_message = self._ws_client.recv() --- modulename: operation, funcname: threading.py(604): with self._cond: ping_background_job.py(66): self._container.ping() --- modulename: reconnecting_websocket, funcname: recv operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) --- modulename: threading, funcname: __exit__ --- modulename: metadata_container, funcname: ping reconnecting_websocket.py(47): if not self.client.connected: --- modulename: operation, funcname: all_subclasses threading.py(268): return self._lock.__exit__(*args) metadata_container.py(256): self._backend.ping(self._id, self.container_type) --- modulename: websocket_client_adapter, funcname: connected operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) websocket_signals_background_job.py(56): self._started = True --- modulename: utils, funcname: wrapper websocket_client_adapter.py(69): return self._ws_client and self._ws_client.connected --- modulename: operation, funcname: backgroud_job_list.py(36): for job in self._jobs: utils.py(58): last_exception = None reconnecting_websocket.py(48): self._try_to_establish_connection() operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) backgroud_job_list.py(37): job.start(container) utils.py(59): start_time = time.monotonic() --- modulename: reconnecting_websocket, funcname: _try_to_establish_connection operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) --- modulename: std_capture_background_job, funcname: start utils.py(60): for retry in itertools.count(0): reconnecting_websocket.py(75): try: --- modulename: operation, funcname: all_subclasses std_capture_background_job.py(39): self._logger = StdoutCaptureLogger(container, self._attribute_name) utils.py(61): if time.monotonic() - start_time > retries_timeout: reconnecting_websocket.py(76): self._request_token_refresh() operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) --- modulename: std_stream_capture_logger, funcname: __init__ utils.py(64): try: --- modulename: reconnecting_websocket, funcname: _request_token_refresh --- modulename: operation, funcname: std_stream_capture_logger.py(59): super().__init__(container, attribute_name, sys.stdout) utils.py(65): return func(*args, **kwargs) reconnecting_websocket.py(88): self._token = self._oauth2_session.refresh_token(token_url=self._oauth2_session.auto_refresh_url) operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) --- modulename: std_stream_capture_logger, funcname: __init__ --- modulename: hosted_neptune_backend, funcname: ping --- modulename: oauth2_session, funcname: refresh_token operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) std_stream_capture_logger.py(29): self._logger = NeptuneLogger(container, attribute_name) hosted_neptune_backend.py(429): "_request_options": { oauth2_session.py(409): if not token_url: --- modulename: operation, funcname: all_subclasses --- modulename: logger, funcname: __init__ hosted_neptune_backend.py(430): "timeout": 10, oauth2_session.py(412): if not is_secure_transport(token_url): operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) logger.py(27): self._container = container hosted_neptune_backend.py(431): "connect_timeout": 10, --- modulename: utils, funcname: is_secure_transport --- modulename: operation, funcname: logger.py(28): self._attribute_name = attribute_name hosted_neptune_backend.py(429): "_request_options": { utils.py(81): if os.environ.get('OAUTHLIB_INSECURE_TRANSPORT'): operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) std_stream_capture_logger.py(30): self.stream = stream hosted_neptune_backend.py(428): request_kwargs = { --- modulename: _collections_abc, funcname: get operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) std_stream_capture_logger.py(31): self._thread_local = threading.local() hosted_neptune_backend.py(434): try: _collections_abc.py(823): try: --- modulename: operation, funcname: all_subclasses std_stream_capture_logger.py(32): self.enabled = True hosted_neptune_backend.py(435): self.leaderboard_client.api.ping( _collections_abc.py(824): return self[key] operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) std_stream_capture_logger.py(33): self._log_data_queue = Queue() --- modulename: swagger_client_wrapper, funcname: __getattr__ --- modulename: os, funcname: __getitem__ --- modulename: operation, funcname: --- modulename: queue, funcname: __init__ swagger_client_wrapper.py(110): return ApiMethodWrapper(getattr(self._api_obj, item)) os.py(676): try: operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) queue.py(35): self.maxsize = maxsize --- modulename: client, funcname: __getattr__ os.py(677): value = self._data[self.encodekey(key)] operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) queue.py(36): self._init(maxsize) client.py(229): return CallableOperation(getattr(self.resource, name), self.also_return_response) --- modulename: os, funcname: encode --- modulename: operation, funcname: all_subclasses --- modulename: queue, funcname: _init --- modulename: resource, funcname: __getattr__ os.py(756): if not isinstance(value, str): operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) queue.py(207): self.queue = deque() resource.py(144): op = self.operations.get(item) os.py(758): return value.encode(encoding, 'surrogateescape') --- modulename: operation, funcname: queue.py(42): self.mutex = threading.Lock() resource.py(145): if not op: os.py(678): except KeyError: operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) queue.py(46): self.not_empty = threading.Condition(self.mutex) resource.py(147): return op os.py(680): raise KeyError(key) from None operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) --- modulename: threading, funcname: __init__ --- modulename: client, funcname: __init__ _collections_abc.py(825): except KeyError: --- modulename: operation, funcname: all_subclasses threading.py(237): if lock is None: client.py(246): self.also_return_response = also_return_response _collections_abc.py(826): return default operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(239): self._lock = lock client.py(247): self.operation = operation utils.py(83): return uri.lower().startswith('https://') --- modulename: operation, funcname: threading.py(241): self.acquire = lock.acquire --- modulename: swagger_client_wrapper, funcname: __init__ oauth2_session.py(415): refresh_token = refresh_token or self.token.get("refresh_token") operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(242): self.release = lock.release swagger_client_wrapper.py(42): self._api_method = api_method --- modulename: oauth2_session, funcname: token operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(246): try: hosted_neptune_backend.py(436): experimentId=container_id, oauth2_session.py(124): return getattr(self._client, "token", None) --- modulename: operation, funcname: all_subclasses threading.py(247): self._release_save = lock._release_save hosted_neptune_backend.py(435): self.leaderboard_client.api.ping( oauth2_session.py(417): log.debug( operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(248): except AttributeError: hosted_neptune_backend.py(437): **request_kwargs, oauth2_session.py(418): "Adding auto refresh key word arguments %s.", self.auto_refresh_kwargs --- modulename: operation, funcname: threading.py(249): pass hosted_neptune_backend.py(435): self.leaderboard_client.api.ping( oauth2_session.py(417): log.debug( operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(250): try: --- modulename: swagger_client_wrapper, funcname: __call__ --- modulename: __init__, funcname: debug operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(251): self._acquire_restore = lock._acquire_restore swagger_client_wrapper.py(95): try: __init__.py(1464): if self.isEnabledFor(DEBUG): --- modulename: operation, funcname: all_subclasses threading.py(252): except AttributeError: swagger_client_wrapper.py(96): future = self._api_method(*args, **kwargs) --- modulename: __init__, funcname: isEnabledFor operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(253): pass --- modulename: client, funcname: __call__ __init__.py(1728): if self.disabled: --- modulename: operation, funcname: threading.py(254): try: client.py(264): log.debug(u'%s(%s)', self.operation.operation_id, op_kwargs) __init__.py(1731): try: operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(255): self._is_owned = lock._is_owned --- modulename: __init__, funcname: debug __init__.py(1732): return self._cache[level] operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(256): except AttributeError: __init__.py(1464): if self.isEnabledFor(DEBUG): oauth2_session.py(420): kwargs.update(self.auto_refresh_kwargs) --- modulename: operation, funcname: all_subclasses threading.py(257): pass --- modulename: __init__, funcname: isEnabledFor oauth2_session.py(421): body = self._client.prepare_refresh_body( operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(258): self._waiters = _deque() __init__.py(1728): if self.disabled: oauth2_session.py(422): body=body, refresh_token=refresh_token, scope=self.scope, **kwargs --- modulename: operation, funcname: queue.py(50): self.not_full = threading.Condition(self.mutex) __init__.py(1731): try: oauth2_session.py(421): body = self._client.prepare_refresh_body( operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) --- modulename: threading, funcname: __init__ __init__.py(1732): return self._cache[level] oauth2_session.py(422): body=body, refresh_token=refresh_token, scope=self.scope, **kwargs operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(237): if lock is None: client.py(265): warn_for_deprecated_op(self.operation) oauth2_session.py(421): body = self._client.prepare_refresh_body( --- modulename: operation, funcname: all_subclasses threading.py(239): self._lock = lock --- modulename: warning, funcname: warn_for_deprecated_op --- modulename: base, funcname: prepare_refresh_body operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(241): self.acquire = lock.acquire warning.py(17): if op.op_spec.get('deprecated', False): base.py(447): refresh_token = refresh_token or self.refresh_token --- modulename: operation, funcname: threading.py(242): self.release = lock.release client.py(268): request_options = op_kwargs.pop('_request_options', {}) base.py(448): scope = self.scope if scope is None else scope operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(246): try: client.py(269): request_config = RequestConfig(request_options, self.also_return_response) base.py(449): return prepare_token_request(self.refresh_token_key, body=body, scope=scope, --- modulename: operation, funcname: all_subclasses threading.py(247): self._release_save = lock._release_save --- modulename: config, funcname: __init__ base.py(450): refresh_token=refresh_token, **kwargs) operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(248): except AttributeError: config.py(81): request_options = request_options.copy() # don't modify the original object base.py(449): return prepare_token_request(self.refresh_token_key, body=body, scope=scope, --- modulename: operation, funcname: threading.py(249): pass config.py(82): self.also_return_response = also_return_response_default base.py(450): refresh_token=refresh_token, **kwargs) operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(250): try: config.py(84): for key in list(request_options.keys()): base.py(449): return prepare_token_request(self.refresh_token_key, body=body, scope=scope, operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(251): self._acquire_restore = lock._acquire_restore config.py(85): if hasattr(self, key): --- modulename: parameters, funcname: prepare_token_request --- modulename: operation, funcname: all_subclasses threading.py(252): except AttributeError: config.py(86): setattr(self, key, request_options.pop(key)) parameters.py(146): params = [('grant_type', grant_type)] operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(253): pass config.py(84): for key in list(request_options.keys()): parameters.py(148): if 'scope' in kwargs: --- modulename: operation, funcname: threading.py(254): try: config.py(85): if hasattr(self, key): parameters.py(149): kwargs['scope'] = list_to_scope(kwargs['scope']) operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(255): self._is_owned = lock._is_owned config.py(86): setattr(self, key, request_options.pop(key)) --- modulename: utils, funcname: list_to_scope operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(256): except AttributeError: config.py(84): for key in list(request_options.keys()): utils.py(16): if isinstance(scope, str) or scope is None: --- modulename: operation, funcname: all_subclasses threading.py(257): pass config.py(88): self.additional_properties = request_options utils.py(17): return scope operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(258): self._waiters = _deque() client.py(271): request_params = construct_request( parameters.py(152): client_id = kwargs.pop('client_id', None) --- modulename: operation, funcname: queue.py(54): self.all_tasks_done = threading.Condition(self.mutex) client.py(272): self.operation, request_options, **op_kwargs) parameters.py(153): if include_client_id: operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) --- modulename: threading, funcname: __init__ client.py(271): request_params = construct_request( parameters.py(154): if client_id is not None: operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(237): if lock is None: client.py(272): self.operation, request_options, **op_kwargs) parameters.py(155): params.append(('client_id', client_id)) operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(239): self._lock = lock client.py(271): request_params = construct_request( parameters.py(158): if code_verifier is not None: operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(241): self.acquire = lock.acquire --- modulename: client, funcname: construct_request parameters.py(163): client_secret = kwargs.pop('client_secret', None) operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(242): self.release = lock.release client.py(294): url = operation.swagger_spec.api_url.rstrip('/') + operation.path_name parameters.py(164): if client_secret is not None: operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(246): try: client.py(296): 'method': str(operation.http_method.upper()), parameters.py(168): for k in kwargs: --- modulename: operation, funcname: all_subclasses threading.py(247): self._release_save = lock._release_save client.py(297): 'url': url, parameters.py(169): if kwargs[k]: operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(248): except AttributeError: client.py(298): 'params': {}, # filled in downstream parameters.py(168): for k in kwargs: --- modulename: operation, funcname: threading.py(249): pass client.py(301): if 'headers' in request_options else {}), parameters.py(169): if kwargs[k]: operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(250): try: client.py(295): request = { parameters.py(170): params.append((str(k), kwargs[k])) operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(251): self._acquire_restore = lock._acquire_restore client.py(304): if request_options.get('use_msgpack', False): parameters.py(168): for k in kwargs: --- modulename: operation, funcname: all_subclasses threading.py(252): except AttributeError: client.py(308): for request_option in ('connect_timeout', 'timeout'): parameters.py(172): return add_params_to_qs(body, params) operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(253): pass client.py(309): if request_option in request_options: --- modulename: common, funcname: add_params_to_qs --- modulename: operation, funcname: threading.py(254): try: client.py(310): request[request_option] = request_options[request_option] common.py(235): if isinstance(params, dict): operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(255): self._is_owned = lock._is_owned client.py(308): for request_option in ('connect_timeout', 'timeout'): common.py(237): queryparams = urlparse.parse_qsl(query, keep_blank_values=True) operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(256): except AttributeError: client.py(309): if request_option in request_options: --- modulename: parse, funcname: parse_qsl --- modulename: operation, funcname: all_subclasses threading.py(257): pass client.py(310): request[request_option] = request_options[request_option] parse.py(742): qs, _coerce_result = _coerce_args(qs) operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(258): self._waiters = _deque() client.py(308): for request_option in ('connect_timeout', 'timeout'): --- modulename: parse, funcname: _coerce_args --- modulename: operation, funcname: queue.py(55): self.unfinished_tasks = 0 client.py(312): construct_params(operation, request, op_kwargs) parse.py(120): str_input = isinstance(args[0], str) operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) std_stream_capture_logger.py(34): self._logging_thread = threading.Thread(target=self.__proces_logs, daemon=True) --- modulename: client, funcname: construct_params parse.py(121): for arg in args[1:]: operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) --- modulename: threading, funcname: __init__ client.py(328): current_params = operation.params.copy() parse.py(126): if str_input: --- modulename: operation, funcname: all_subclasses threading.py(850): assert group is None, "group argument must be None for now" --- modulename: util, funcname: copy parse.py(127): return args + (_noop,) operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(851): if kwargs is None: util.py(196): copied_dict = type(self)(self) parse.py(743): separator, _ = _coerce_args(separator) --- modulename: operation, funcname: threading.py(852): kwargs = {} --- modulename: util, funcname: __init__ --- modulename: parse, funcname: _coerce_args operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(853): if name: util.py(157): super(AliasKeyDict, self).__init__(*args, **kwargs) parse.py(120): str_input = isinstance(args[0], str) operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(856): name = _newname("Thread-%d") util.py(158): self.alias_to_key = {} # type: typing.Dict[typing.Text, typing.Any] parse.py(121): for arg in args[1:]: --- modulename: operation, funcname: all_subclasses --- modulename: threading, funcname: _newname util.py(197): copied_dict.alias_to_key = self.alias_to_key.copy() parse.py(126): if str_input: operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(783): return name_template % _counter() util.py(198): return copied_dict parse.py(127): return args + (_noop,) --- modulename: operation, funcname: threading.py(857): if target is not None: client.py(329): for param_name, param_value in iteritems(op_kwargs): parse.py(745): if not separator or (not isinstance(separator, (str, bytes))): operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(858): try: --- modulename: six, funcname: iteritems parse.py(751): if max_num_fields is not None: operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(859): target_name = target.__name__ six.py(605): return iter(d.items(**kw)) parse.py(756): r = [] --- modulename: operation, funcname: all_subclasses threading.py(860): name += f" ({target_name})" client.py(330): param = current_params.pop(param_name, None) parse.py(757): for name_value in qs.split(separator): operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(864): self._target = target --- modulename: util, funcname: pop parse.py(758): if not name_value and not strict_parsing: --- modulename: operation, funcname: threading.py(865): self._name = name util.py(177): return super(AliasKeyDict, self).pop(self.determine_key(key), default) parse.py(759): continue operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(866): self._args = args --- modulename: util, funcname: determine_key parse.py(757): for name_value in qs.split(separator): operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(867): self._kwargs = kwargs util.py(167): if key in self.alias_to_key: # this will normally be False, optimize for it parse.py(777): return r --- modulename: operation, funcname: all_subclasses threading.py(868): if daemon is not None: util.py(169): return key common.py(238): queryparams.extend(params) operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(869): self._daemonic = daemon client.py(331): if param is None: common.py(239): return urlencode(queryparams) --- modulename: operation, funcname: threading.py(872): self._ident = None client.py(335): marshal_param(param, param_value, request) --- modulename: common, funcname: urlencode operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(873): if _HAVE_THREAD_NATIVE_ID: --- modulename: param, funcname: marshal_param common.py(64): utf8_params = encode_params_utf8(params) operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(874): self._native_id = None param.py(119): swagger_spec = param.swagger_spec --- modulename: common, funcname: encode_params_utf8 --- modulename: operation, funcname: all_subclasses threading.py(875): self._tstate_lock = None param.py(120): deref = swagger_spec.deref common.py(76): encoded = [] operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(876): self._started = Event() param.py(122): param_spec = deref(get_param_type_spec(param)) common.py(77): for k, v in params: --- modulename: operation, funcname: --- modulename: threading, funcname: __init__ --- modulename: param, funcname: get_param_type_spec common.py(78): encoded.append(( operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(546): self._cond = Condition(Lock()) param.py(94): location = param.location common.py(79): k.encode('utf-8') if isinstance(k, str) else k, operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) --- modulename: threading, funcname: __init__ --- modulename: param, funcname: location common.py(80): v.encode('utf-8') if isinstance(v, str) else v)) --- modulename: operation, funcname: all_subclasses threading.py(237): if lock is None: param.py(65): return self.param_spec['in'] common.py(78): encoded.append(( operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(239): self._lock = lock param.py(95): if location in ('path', 'query', 'header', 'formData'): common.py(77): for k, v in params: --- modulename: operation, funcname: threading.py(241): self.acquire = lock.acquire param.py(96): return param.param_spec common.py(78): encoded.append(( operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(242): self.release = lock.release --- modulename: spec, funcname: _force_deref common.py(79): k.encode('utf-8') if isinstance(k, str) else k, operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(246): try: spec.py(407): if ref_dict is None or not is_ref(ref_dict): common.py(80): v.encode('utf-8') if isinstance(v, str) else v)) --- modulename: operation, funcname: all_subclasses threading.py(247): self._release_save = lock._release_save --- modulename: schema, funcname: is_ref common.py(78): encoded.append(( operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(248): except AttributeError: schema.py(74): try: common.py(77): for k, v in params: --- modulename: operation, funcname: threading.py(249): pass schema.py(75): return '$ref' in spec and is_dict_like(spec) and isinstance(spec['$ref'], string_types) common.py(78): encoded.append(( operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(250): try: spec.py(408): return ref_dict common.py(79): k.encode('utf-8') if isinstance(k, str) else k, operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(251): self._acquire_restore = lock._acquire_restore param.py(123): location = param.location common.py(80): v.encode('utf-8') if isinstance(v, str) else v)) --- modulename: operation, funcname: all_subclasses threading.py(252): except AttributeError: --- modulename: param, funcname: location common.py(78): encoded.append(( operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(253): pass param.py(65): return self.param_spec['in'] common.py(77): for k, v in params: --- modulename: operation, funcname: threading.py(254): try: param.py(127): if value is None and not param.required: common.py(81): return encoded operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(255): self._is_owned = lock._is_owned param.py(130): value = marshal_schema_object(swagger_spec, param_spec, value) common.py(65): urlencoded = _urlencode(utf8_params) operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(256): except AttributeError: --- modulename: marshal, funcname: marshal_schema_object --- modulename: parse, funcname: urlencode --- modulename: operation, funcname: all_subclasses threading.py(257): pass marshal.py(53): marshaling_method = _get_marshaling_method(swagger_spec=swagger_spec, object_schema=schema_object_spec) parse.py(927): if hasattr(query, "items"): operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(258): self._waiters = _deque() --- modulename: _decorators, funcname: wrapper parse.py(932): try: --- modulename: operation, funcname: threading.py(547): self._flag = False _decorators.py(80): try: parse.py(935): if len(query) and not isinstance(query[0], tuple): operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(877): self._is_stopped = False _decorators.py(81): return func(*args, **kwargs) parse.py(946): l = [] operation.py(51): return set(cls.__subclasses__()).union([s for c in cls.__subclasses__() for s in all_subclasses(c)]) threading.py(878): self._initialized = True --- modulename: util, funcname: wrapper parse.py(947): if not doseq: --- modulename: operation, funcname: threading.py(880): self._stderr = _sys.stderr util.py(119): cache_key = make_key(*args, **kwargs) parse.py(948): for k, v in query: operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} threading.py(881): self._invoke_excepthook = _make_invoke_excepthook() --- modulename: util, funcname: make_key parse.py(949): if isinstance(k, bytes): operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} --- modulename: threading, funcname: _make_invoke_excepthook util.py(105): if args: parse.py(950): k = quote_via(k, safe) operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} threading.py(1306): old_excepthook = excepthook util.py(109): param_name_to_value_mapping = sorted(iteritems(dict(default_mapping, **kwargs))) --- modulename: parse, funcname: quote_plus operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} threading.py(1307): old_sys_excepthook = _sys.excepthook --- modulename: six, funcname: iteritems parse.py(878): if ((isinstance(string, str) and ' ' not in string) or operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} threading.py(1308): if old_excepthook is None: six.py(605): return iter(d.items(**kw)) parse.py(879): (isinstance(string, bytes) and b' ' not in string)): operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} threading.py(1310): if old_sys_excepthook is None: util.py(111): return tuple( parse.py(878): if ((isinstance(string, str) and ' ' not in string) or operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} threading.py(1313): sys_exc_info = _sys.exc_info util.py(113): for param_name, param_value in param_name_to_value_mapping parse.py(879): (isinstance(string, bytes) and b' ' not in string)): operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} threading.py(1314): local_print = print util.py(111): return tuple( parse.py(880): return quote(string, safe, encoding, errors) operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} threading.py(1315): local_sys = _sys --- modulename: util, funcname: --- modulename: parse, funcname: quote operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} threading.py(1317): def invoke_excepthook(thread): util.py(111): return tuple( parse.py(856): if isinstance(string, str): operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} threading.py(1349): return invoke_excepthook util.py(113): for param_name, param_value in param_name_to_value_mapping parse.py(865): if encoding is not None: operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} threading.py(883): _dangling.add(self) util.py(112): (param_name, id(param_value)) parse.py(867): if errors is not None: operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} --- modulename: _weakrefset, funcname: add util.py(111): return tuple( parse.py(869): return quote_from_bytes(string, safe) operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} _weakrefset.py(87): if self._pending_removals: --- modulename: util, funcname: --- modulename: parse, funcname: quote_from_bytes operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} _weakrefset.py(89): self.data.add(ref(item, self._remove)) util.py(111): return tuple( parse.py(893): if not isinstance(bs, (bytes, bytearray)): operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} std_stream_capture_logger.py(35): self._logging_thread.start() util.py(113): for param_name, param_value in param_name_to_value_mapping parse.py(895): if not bs: operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} --- modulename: threading, funcname: start util.py(112): (param_name, id(param_value)) parse.py(897): if isinstance(safe, str): operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} threading.py(926): if not self._initialized: util.py(111): return tuple( parse.py(899): safe = safe.encode('ascii', 'ignore') operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} threading.py(929): if self._started.is_set(): --- modulename: util, funcname: parse.py(902): if not bs.rstrip(_ALWAYS_SAFE_BYTES + safe): operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} --- modulename: threading, funcname: is_set util.py(111): return tuple( parse.py(903): return bs.decode() operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} threading.py(555): return self._flag util.py(113): for param_name, param_value in param_name_to_value_mapping parse.py(954): if isinstance(v, bytes): operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} threading.py(932): with _active_limbo_lock: util.py(112): (param_name, id(param_value)) parse.py(955): v = quote_via(v, safe) operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} threading.py(933): _limbo[self] = self util.py(111): return tuple( --- modulename: parse, funcname: quote_plus operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} threading.py(932): with _active_limbo_lock: --- modulename: util, funcname: parse.py(878): if ((isinstance(string, str) and ' ' not in string) or operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} threading.py(934): try: util.py(111): return tuple( parse.py(879): (isinstance(string, bytes) and b' ' not in string)): operation.py(73): sub_classes = {cls.__name__: cls for cls in all_subclasses(Operation)} threading.py(935): _start_new_thread(self._bootstrap, ()) util.py(120): cached_value = cache.get(cache_key, _CACHE_MISS) parse.py(878): if ((isinstance(string, str) and ' ' not in string) or operation.py(74): if not data["type"] in sub_classes: util.py(121): if cached_value is _CACHE_MISS: threading.py(940): self._started.wait() operation.py(76): return sub_classes[data["type"]].from_dict(data) util.py(122): if cache_key in key_in_progress_set: --- modulename: threading, funcname: run parse.py(879): (isinstance(string, bytes) and b' ' not in string)): util.py(124): key_in_progress_set.add(cache_key) threading.py(951): try: --- modulename: threading, funcname: wait util.py(125): cached_value = func(*args, **kwargs) threading.py(952): if self._target is not None: --- modulename: operation, funcname: from_dict parse.py(880): return quote(string, safe, encoding, errors) threading.py(953): self._target(*self._args, **self._kwargs) operation.py(148): return AssignString(data["path"], data["value"]) threading.py(604): with self._cond: --- modulename: std_stream_capture_logger, funcname: __proces_logs --- modulename: operation, funcname: __init__ --- modulename: threading, funcname: __enter__ std_stream_capture_logger.py(50): while True: (3): (4): disk_queue.py(139): return QueueElement[T](obj, ver, size) --- modulename: parse, funcname: quote std_stream_capture_logger.py(51): data = self._log_data_queue.get() --- modulename: typing, funcname: inner parse.py(856): if isinstance(string, str): --- modulename: queue, funcname: get typing.py(308): try: threading.py(265): return self._lock.__enter__() queue.py(165): with self.not_empty: typing.py(309): return cached(*args, **kwds) threading.py(605): signaled = self._flag --- modulename: threading, funcname: __enter__ --- modulename: typing, funcname: __call__ parse.py(865): if encoding is not None: threading.py(265): return self._lock.__enter__() threading.py(606): if not signaled: parse.py(867): if errors is not None: typing.py(954): if not self._inst: threading.py(608): return signaled parse.py(869): return quote_from_bytes(string, safe) --- modulename: marshal, funcname: _get_marshaling_method threading.py(604): with self._cond: --- modulename: parse, funcname: quote_from_bytes queue.py(166): if not block: --- modulename: threading, funcname: __exit__ parse.py(893): if not isinstance(bs, (bytes, bytearray)): typing.py(957): result = self.__origin__(*args, **kwargs) threading.py(268): return self._lock.__exit__(*args) parse.py(895): if not bs: marshal.py(165): object_schema = swagger_spec.deref(object_schema) std_stream_capture_logger.py(60): sys.stdout = self parse.py(897): if isinstance(safe, str): queue.py(169): elif timeout is None: backgroud_job_list.py(36): for job in self._jobs: parse.py(899): safe = safe.encode('ascii', 'ignore') --- modulename: disk_queue, funcname: __init__ queue.py(170): while not self._qsize(): --- modulename: spec, funcname: _force_deref (3): spec.py(407): if ref_dict is None or not is_ref(ref_dict): ^CTraceback (most recent call last): File "/home/tyler/opt/anaconda/envs/gaddy/lib/python3.10/runpy.py", line 196, in _run_module_as_main return _run_code(code, main_globals, None, File "/home/tyler/opt/anaconda/envs/gaddy/lib/python3.10/runpy.py", line 86, in _run_code exec(code, run_globals) File "/home/tyler/opt/anaconda/envs/gaddy/lib/python3.10/trace.py", line 740, in main() File "/home/tyler/opt/anaconda/envs/gaddy/lib/python3.10/trace.py", line 728, in main t.runctx(code, globs, globs) File "/home/tyler/opt/anaconda/envs/gaddy/lib/python3.10/trace.py", line 450, in runctx exec(cmd, globals, locals) File "2023-06-22_librispeech_contrastive.py", line 756, in neptune_logger.log_hyperparams(vars(config)) File "/home/tyler/opt/anaconda/envs/gaddy/lib/python3.10/site-packages/lightning_utilities/core/rank_zero.py", line 27, in wrapped_fn return fn(*args, **kwargs) File "/home/tyler/opt/anaconda/envs/gaddy/lib/python3.10/site-packages/pytorch_lightning/loggers/neptune.py", line 406, in log_hyperparams self.run[parameters_key] = params File "/home/tyler/opt/anaconda/envs/gaddy/lib/python3.10/site-packages/lightning_fabric/loggers/logger.py", line 114, in experiment return fn(self) File "/home/tyler/opt/anaconda/envs/gaddy/lib/python3.10/site-packages/pytorch_lightning/loggers/neptune.py", line 358, in run self._run_instance = neptune.init_run(**self._neptune_init_args) File "/home/tyler/opt/anaconda/envs/gaddy/lib/python3.10/site-packages/neptune/metadata_containers/run.py", line 364, in __init__ super().__init__(project=project, api_token=api_token, mode=mode, flush_period=flush_period, proxies=proxies) File "/home/tyler/opt/anaconda/envs/gaddy/lib/python3.10/site-packages/neptune/metadata_containers/metadata_container.py", line 144, in __init__ self._startup(debug_mode=mode == Mode.DEBUG) File "/home/tyler/opt/anaconda/envs/gaddy/lib/python3.10/site-packages/neptune/metadata_containers/metadata_container.py", line 486, in _startup self.start() File "/home/tyler/opt/anaconda/envs/gaddy/lib/python3.10/site-packages/neptune/metadata_containers/metadata_container.py", line 261, in start self._bg_job.start(self) File "/home/tyler/opt/anaconda/envs/gaddy/lib/python3.10/site-packages/neptune/internal/backgroud_job_list.py", line 36, in start for job in self._jobs: File "/home/tyler/opt/anaconda/envs/gaddy/lib/python3.10/site-packages/neptune/internal/backgroud_job_list.py", line 36, in start for job in self._jobs: File "/home/tyler/opt/anaconda/envs/gaddy/lib/python3.10/trace.py", line 575, in localtrace_trace print("%s(%d): %s" % (bname, lineno, File "/home/tyler/opt/anaconda/envs/gaddy/lib/python3.10/site-packages/neptune/internal/streams/std_stream_capture_logger.py", line 39, in write self._log_data_queue.put_nowait(data) File "/home/tyler/opt/anaconda/envs/gaddy/lib/python3.10/queue.py", line 191, in put_nowait return self.put(item, block=False) File "/home/tyler/opt/anaconda/envs/gaddy/lib/python3.10/queue.py", line 133, in put with self.not_full: File "/home/tyler/opt/anaconda/envs/gaddy/lib/python3.10/threading.py", line 265, in __enter__ return self._lock.__enter__() KeyboardInterrupt [1] + 51627 killed python train.py