From a23d0d8617ba3119069e610fc7b0850a17322726 Mon Sep 17 00:00:00 2001 From: Craig Northway Date: Fri, 25 Jul 2014 11:26:48 +1000 Subject: [PATCH] Autopep8 fixes with maximum line length 120 --- git/__init__.py | 23 +- git/base.py | 215 ++++---- git/cmd.py | 246 ++++----- git/config.py | 205 ++++---- git/db/cmd/base.py | 395 +++++++------- git/db/cmd/complex.py | 6 +- git/db/compat.py | 17 +- git/db/complex.py | 8 + git/db/dulwich/__init__.py | 4 +- git/db/dulwich/complex.py | 57 +- git/db/interface.py | 328 ++++++------ git/db/py/base.py | 278 +++++----- git/db/py/complex.py | 78 +-- git/db/py/loose.py | 128 +++-- git/db/py/mem.py | 65 +-- git/db/py/pack.py | 88 ++-- git/db/py/ref.py | 20 +- git/db/py/resolve.py | 162 +++--- git/db/py/submodule.py | 13 +- git/db/py/transport.py | 35 +- git/db/pygit2/__init__.py | 3 +- git/db/pygit2/complex.py | 59 ++- git/diff.py | 141 ++--- git/exc.py | 28 +- git/fun.py | 276 +++++----- git/index/__init__.py | 2 +- git/index/base.py | 222 ++++---- git/index/fun.py | 131 ++--- git/index/typ.py | 34 +- git/index/util.py | 19 +- git/objects/__init__.py | 6 +- git/objects/base.py | 87 +-- git/objects/blob.py | 4 +- git/objects/commit.py | 189 ++++--- git/objects/fun.py | 73 +-- git/objects/submodule/base.py | 380 ++++++------- git/objects/submodule/root.py | 215 ++++---- git/objects/submodule/util.py | 33 +- git/objects/tag.py | 41 +- git/objects/tree.py | 126 ++--- git/objects/util.py | 161 +++--- git/odict.py | 419 ++++++++++----- git/pack.py | 497 +++++++++--------- git/refs/__init__.py | 2 +- git/refs/head.py | 42 +- git/refs/headref.py | 70 ++- git/refs/log.py | 159 +++--- git/refs/reference.py | 56 +- git/refs/remote.py | 24 +- git/refs/symbolic.py | 305 +++++------ git/refs/tag.py | 38 +- git/remote.py | 140 ++--- git/repo.py | 15 +- git/stream.py | 348 ++++++------ git/test/__init__.py | 2 +- git/test/db/base.py | 285 +++++----- git/test/db/cmd/test_base.py | 66 ++- git/test/db/dulwich/lib.py | 8 +- git/test/db/dulwich/test_base.py | 10 +- git/test/db/lib.py | 105 ++-- git/test/db/py/test_base.py | 6 +- git/test/db/py/test_git.py | 22 +- git/test/db/py/test_loose.py | 18 +- git/test/db/py/test_mem.py | 13 +- git/test/db/py/test_pack.py | 36 +- git/test/db/py/test_ref.py | 32 +- git/test/db/pygit2/lib.py | 8 +- git/test/db/pygit2/test_base.py | 10 +- git/test/db/test_base.py | 2 +- git/test/lib/__init__.py | 6 +- git/test/lib/asserts.py | 18 +- git/test/lib/base.py | 76 +-- git/test/lib/helper.py | 139 ++--- git/test/objects/__init__.py | 1 - git/test/objects/lib.py | 16 +- git/test/objects/test_blob.py | 10 +- git/test/objects/test_commit.py | 132 ++--- git/test/objects/test_submodule.py | 253 +++++---- git/test/objects/test_tree.py | 75 ++- git/test/performance/db/__init__.py | 1 - git/test/performance/db/looseodb_impl.py | 65 +-- git/test/performance/db/odb_impl.py | 34 +- git/test/performance/db/packedodb_impl.py | 38 +- git/test/performance/db/test_looseodb_cmd.py | 7 +- .../performance/db/test_looseodb_dulwich.py | 4 +- git/test/performance/db/test_looseodb_pure.py | 2 +- .../performance/db/test_looseodb_pygit2.py | 4 +- git/test/performance/db/test_odb_cmd.py | 2 +- git/test/performance/db/test_odb_dulwich.py | 4 +- git/test/performance/db/test_odb_pure.py | 2 +- git/test/performance/db/test_odb_pygit2.py | 4 +- .../performance/db/test_packedodb_pure.py | 37 +- git/test/performance/lib.py | 32 +- git/test/performance/objects/__init__.py | 1 - git/test/performance/objects/test_commit.py | 47 +- git/test/performance/test_utils.py | 87 +-- git/test/refs/__init__.py | 1 - git/test/refs/test_reflog.py | 50 +- git/test/refs/test_refs.py | 220 ++++---- git/test/test_base.py | 123 +++-- git/test/test_cmd.py | 37 +- git/test/test_config.py | 49 +- git/test/test_diff.py | 57 +- git/test/test_example.py | 28 +- git/test/test_fun.py | 114 ++-- git/test/test_import.py | 27 +- git/test/test_index.py | 314 ++++++----- git/test/test_pack.py | 134 +++-- git/test/test_remote.py | 228 ++++---- git/test/test_stats.py | 17 +- git/test/test_stream.py | 78 ++- git/test/test_util.py | 111 ++-- git/typ.py | 4 +- git/util.py | 341 ++++++------ setup.py | 46 +- 115 files changed, 5387 insertions(+), 5028 deletions(-) diff --git a/git/__init__.py b/git/__init__.py index 91a10bb3f..18c84bddd 100644 --- a/git/__init__.py +++ b/git/__init__.py @@ -14,16 +14,16 @@ #{ Initialization def _init_externals(): """Initialize external projects by putting them into the path""" - ext_base = os.path.join(os.path.dirname(__file__), 'ext') + ext_base = os.path.join(os.path.dirname(__file__), 'ext') for package in ('async', 'smmap'): sys.path.append(os.path.join(ext_base, package)) try: __import__(package) except ImportError: raise ImportError("%r could not be found in your PYTHONPATH" % package) - #END verify import - #END handle external import - + # END verify import + # END handle external import + #} END initialization ################# @@ -43,14 +43,13 @@ def _init_externals(): from git.index import * from git.repo import Repo from git.util import ( - LockFile, - BlockingLockFile, - Stats, - Actor - ) + LockFile, + BlockingLockFile, + Stats, + Actor +) #} END imports -__all__ = [ name for name, obj in locals().items() - if not (name.startswith('_') or inspect.ismodule(obj)) ] - +__all__ = [name for name, obj in locals().items() + if not (name.startswith('_') or inspect.ismodule(obj))] diff --git a/git/base.py b/git/base.py index bad5f7472..a3971ce4f 100644 --- a/git/base.py +++ b/git/base.py @@ -4,192 +4,198 @@ # the New BSD License: http://www.opensource.org/licenses/bsd-license.php """Module with basic data structures - they are designed to be lightweight and fast""" from util import ( - bin_to_hex, - zlib - ) + bin_to_hex, + zlib +) from fun import ( - type_id_to_type_map, - type_to_type_id_map - ) + type_id_to_type_map, + type_to_type_id_map +) -__all__ = ('OInfo', 'OPackInfo', 'ODeltaPackInfo', - 'OStream', 'OPackStream', 'ODeltaPackStream', - 'IStream', 'InvalidOInfo', 'InvalidOStream' ) +__all__ = ('OInfo', 'OPackInfo', 'ODeltaPackInfo', + 'OStream', 'OPackStream', 'ODeltaPackStream', + 'IStream', 'InvalidOInfo', 'InvalidOStream') #{ ODB Bases + class OInfo(tuple): + """Carries information about an object in an ODB, provding information about the binary sha of the object, the type_string as well as the uncompressed size in bytes. - + It can be accessed using tuple notation and using attribute access notation:: - + assert dbi[0] == dbi.binsha assert dbi[1] == dbi.type assert dbi[2] == dbi.size - + The type is designed to be as lighteight as possible.""" __slots__ = tuple() - + def __new__(cls, sha, type, size): return tuple.__new__(cls, (sha, type, size)) - + def __init__(self, *args): tuple.__init__(self) - - #{ Interface + + #{ Interface @property def binsha(self): """:return: our sha as binary, 20 bytes""" return self[0] - + @property def hexsha(self): """:return: our sha, hex encoded, 40 bytes""" return bin_to_hex(self[0]) - + @property def type(self): return self[1] - + @property def type_id(self): return type_to_type_id_map[self[1]] - + @property def size(self): return self[2] #} END interface - - + + class OPackInfo(tuple): + """As OInfo, but provides a type_id property to retrieve the numerical type id, and does not include a sha. - + Additionally, the pack_offset is the absolute offset into the packfile at which all object information is located. The data_offset property points to the abosolute location in the pack at which that actual data stream can be found.""" __slots__ = tuple() - + def __new__(cls, packoffset, type, size): - return tuple.__new__(cls, (packoffset,type, size)) - + return tuple.__new__(cls, (packoffset, type, size)) + def __init__(self, *args): tuple.__init__(self) - - #{ Interface - + + #{ Interface + @property def pack_offset(self): return self[0] - + @property def type(self): return type_id_to_type_map[self[1]] - + @property def type_id(self): return self[1] - + @property def size(self): return self[2] - + #} END interface - - + + class ODeltaPackInfo(OPackInfo): + """Adds delta specific information, Either the 20 byte sha which points to some object in the database, or the negative offset from the pack_offset, so that pack_offset - delta_info yields the pack offset of the base object""" __slots__ = tuple() - + def __new__(cls, packoffset, type, size, delta_info): return tuple.__new__(cls, (packoffset, type, size, delta_info)) - - #{ Interface + + #{ Interface @property def delta_info(self): return self[3] - #} END interface - - + #} END interface + + class OStream(OInfo): + """Base for object streams retrieved from the database, providing additional information about the stream. Generally, ODB streams are read-only as objects are immutable""" __slots__ = tuple() - + def __new__(cls, sha, type, size, stream, *args, **kwargs): """Helps with the initialization of subclasses""" return tuple.__new__(cls, (sha, type, size, stream)) - - + def __init__(self, *args, **kwargs): tuple.__init__(self) - - #{ Stream Reader Interface - + + #{ Stream Reader Interface + def read(self, size=-1): return self[3].read(size) - + @property def stream(self): return self[3] - + #} END stream reader interface - - + + class ODeltaStream(OStream): + """Uses size info of its stream, delaying reads""" - + def __new__(cls, sha, type, size, stream, *args, **kwargs): """Helps with the initialization of subclasses""" return tuple.__new__(cls, (sha, type, size, stream)) - + #{ Stream Reader Interface - + @property def size(self): return self[3].size - + #} END stream reader interface - - + + class OPackStream(OPackInfo): + """Next to pack object information, a stream outputting an undeltified base object is provided""" __slots__ = tuple() - + def __new__(cls, packoffset, type, size, stream, *args): """Helps with the initialization of subclasses""" return tuple.__new__(cls, (packoffset, type, size, stream)) - - #{ Stream Reader Interface + + #{ Stream Reader Interface def read(self, size=-1): return self[3].read(size) - + @property def stream(self): return self[3] #} END stream reader interface - + class ODeltaPackStream(ODeltaPackInfo): + """Provides a stream outputting the uncompressed offset delta information""" __slots__ = tuple() - + def __new__(cls, packoffset, type, size, delta_info, stream): return tuple.__new__(cls, (packoffset, type, size, delta_info, stream)) - - #{ Stream Reader Interface + #{ Stream Reader Interface def read(self, size=-1): return self[4].read(size) - + @property def stream(self): return self[4] @@ -197,106 +203,107 @@ def stream(self): class IStream(list): + """Represents an input content stream to be fed into the ODB. It is mutable to allow the ODB to record information about the operations outcome right in this instance. - + It provides interfaces for the OStream and a StreamReader to allow the instance to blend in without prior conversion. - + The only method your content stream must support is 'read'""" __slots__ = tuple() - + def __new__(cls, type, size, stream, sha=None): return list.__new__(cls, (sha, type, size, stream, None)) - + def __init__(self, type, size, stream, sha=None): list.__init__(self, (sha, type, size, stream, None)) - - #{ Interface + + #{ Interface @property def hexsha(self): """:return: our sha, hex encoded, 40 bytes""" return bin_to_hex(self[0]) - + def _error(self): """:return: the error that occurred when processing the stream, or None""" return self[4] - + def _set_error(self, exc): """Set this input stream to the given exc, may be None to reset the error""" self[4] = exc - + error = property(_error, _set_error) - + #} END interface - + #{ Stream Reader Interface - + def read(self, size=-1): """Implements a simple stream reader interface, passing the read call on to our internal stream""" return self[3].read(size) - - #} END stream reader interface - + + #} END stream reader interface + #{ interface - + def _set_binsha(self, binsha): self[0] = binsha - + def _binsha(self): return self[0] - + binsha = property(_binsha, _set_binsha) - - + def _type(self): return self[1] - + def _set_type(self, type): self[1] = type - + type = property(_type, _set_type) - + def _size(self): return self[2] - + def _set_size(self, size): self[2] = size - + size = property(_size, _set_size) - + def _stream(self): return self[3] - + def _set_stream(self, stream): self[3] = stream - + stream = property(_stream, _set_stream) - - #} END odb info interface - + + #} END odb info interface + class InvalidOInfo(tuple): + """Carries information about a sha identifying an object which is invalid in the queried database. The exception attribute provides more information about the cause of the issue""" __slots__ = tuple() - + def __new__(cls, sha, exc): return tuple.__new__(cls, (sha, exc)) - + def __init__(self, sha, exc): tuple.__init__(self, (sha, exc)) - + @property def binsha(self): return self[0] - + @property def hexsha(self): return bin_to_hex(self[0]) - + @property def error(self): """:return: exception instance explaining the failure""" @@ -304,8 +311,8 @@ def error(self): class InvalidOStream(InvalidOInfo): + """Carries information about an invalid ODB stream""" __slots__ = tuple() - -#} END ODB Bases +#} END ODB Bases diff --git a/git/cmd.py b/git/cmd.py index a81919e1f..b0e6b0f49 100644 --- a/git/cmd.py +++ b/git/cmd.py @@ -4,74 +4,77 @@ # This module is part of GitPython and is released under # the BSD License: http://www.opensource.org/licenses/bsd-license.php -import os, sys +import os +import sys from util import ( - LazyMixin, - stream_copy - ) + LazyMixin, + stream_copy +) from exc import GitCommandError from subprocess import ( - call, - Popen, - PIPE - ) + call, + Popen, + PIPE +) execute_kwargs = ('istream', 'with_keep_cwd', 'with_extended_output', - 'with_exceptions', 'as_process', - 'output_stream', 'output_strip' ) + 'with_exceptions', 'as_process', + 'output_stream', 'output_strip') __all__ = ('Git', ) + def dashify(string): return string.replace('_', '-') class Git(LazyMixin): + """ The Git class manages communication with the Git binary. - + It provides a convenient interface to calling the Git binary, such as in:: - + g = Git( git_dir ) g.init() # calls 'git init' program rval = g.ls_files() # calls 'git ls-files' program - + ``Debugging`` Set the GIT_PYTHON_TRACE environment variable print each invocation of the command to stdout. Set its value to 'full' to see details about the returned values. """ __slots__ = ("_working_dir", "cat_file_all", "cat_file_header", "_version_info") - + # CONFIGURATION # The size in bytes read from stdout when copying git's output to another stream - max_chunk_size = 1024*64 - + max_chunk_size = 1024 * 64 + git_exec_name = "git" # default that should work on linux and windows git_exec_name_win = "git.cmd" # alternate command name, windows only - + # Enables debugging of GitPython's git commands GIT_PYTHON_TRACE = os.environ.get("GIT_PYTHON_TRACE", False) - + # Provide the full path to the git executable. Otherwise it assumes git is in the path _git_exec_env_var = "GIT_PYTHON_GIT_EXECUTABLE" GIT_PYTHON_GIT_EXECUTABLE = os.environ.get(_git_exec_env_var, git_exec_name) - - + class AutoInterrupt(object): + """Kill/Interrupt the stored process instance once this instance goes out of scope. It is used to prevent processes piling up in case iterators stop reading. Besides all attributes are wired through to the contained process object. - + The wait method was overridden to perform automatic status code checking and possibly raise.""" - __slots__= ("proc", "args") - - def __init__(self, proc, args ): + __slots__ = ("proc", "args") + + def __init__(self, proc, args): self.proc = proc self.args = args - + def __del__(self): self.proc.stdout.close() self.proc.stderr.close() @@ -79,61 +82,62 @@ def __del__(self): # did the process finish already so we have a return code ? if self.proc.poll() is not None: return - - # can be that nothing really exists anymore ... + + # can be that nothing really exists anymore ... if os is None: return - + # try to kill it try: os.kill(self.proc.pid, 2) # interrupt signal except OSError: - pass # ignore error when process already died + pass # ignore error when process already died except AttributeError: - # try windows - # for some reason, providing None for stdout/stderr still prints something. This is why - # we simply use the shell and redirect to nul. Its slower than CreateProcess, question + # try windows + # for some reason, providing None for stdout/stderr still prints something. This is why + # we simply use the shell and redirect to nul. Its slower than CreateProcess, question # is whether we really want to see all these messages. Its annoying no matter what. call(("TASKKILL /F /T /PID %s 2>nul 1>nul" % str(self.proc.pid)), shell=True) - # END exception handling - + # END exception handling + def __getattr__(self, attr): return getattr(self.proc, attr) - + def wait(self): """Wait for the process and return its status code. - + :raise GitCommandError: if the return status is not 0""" status = self.proc.wait() self.proc.stdout.close() self.proc.stderr.close() if status != 0: raise GitCommandError(self.args, status, self.proc.stderr.read()) - # END status handling + # END status handling return status # END auto interrupt - + class CatFileContentStream(object): + """Object representing a sized read-only stream returning the contents of an object. It behaves like a stream, but counts the data read and simulates an empty stream once our sized content region is empty. If not all data is read to the end of the objects's lifetime, we read the rest to assure the underlying stream continues to work""" - + __slots__ = ('_stream', '_nbr', '_size') - + def __init__(self, size, stream): self._stream = stream self._size = size self._nbr = 0 # num bytes read - - # special case: if the object is empty, has null bytes, get the + + # special case: if the object is empty, has null bytes, get the # final newline right away. if size == 0: stream.read(1) # END handle empty streams - + def read(self, size=-1): bytes_left = self._size - self._nbr if bytes_left == 0: @@ -147,17 +151,17 @@ def read(self, size=-1): # END check early depletion data = self._stream.read(size) self._nbr += len(data) - + # check for depletion, read our final byte to make the stream usable by others if self._size - self._nbr == 0: self._stream.read(1) # final newline # END finish reading return data - + def readline(self, size=-1): if self._nbr == self._size: return '' - + # clamp size to lowest allowed value bytes_left = self._size - self._nbr if size > -1: @@ -165,21 +169,21 @@ def readline(self, size=-1): else: size = bytes_left # END handle size - + data = self._stream.readline(size) self._nbr += len(data) - + # handle final byte if self._size - self._nbr == 0: self._stream.read(1) # END finish reading - + return data - + def readlines(self, size=-1): if self._nbr == self._size: return list() - + # leave all additional logic to our readline method, we just check the size out = list() nbr = 0 @@ -195,16 +199,16 @@ def readlines(self, size=-1): # END handle size constraint # END readline loop return out - + def __iter__(self): return self - + def next(self): line = self.readline() if not line: raise StopIteration return line - + def __del__(self): bytes_left = self._size - self._nbr if bytes_left: @@ -212,11 +216,10 @@ def __del__(self): # includes terminating newline self._stream.read(bytes_left + 1) # END handle incomplete read - - + def __init__(self, working_dir=None): """Initialize this instance with: - + :param working_dir: Git directory we should work in. If None, we always work in the current directory as returned by os.getcwd(). @@ -224,7 +227,7 @@ def __init__(self, working_dir=None): .git directory in case of bare repositories.""" super(Git, self).__init__() self._working_dir = working_dir - + # cached command slots self.cat_file_header = None self.cat_file_all = None @@ -244,14 +247,13 @@ def _set_cache_(self, attr): self._version_info = tuple(int(n) for n in version_numbers.split('.')[:4]) else: super(Git, self)._set_cache_(attr) - #END handle version info - + # END handle version info @property def working_dir(self): """:return: Git directory we are working on""" return self._working_dir - + @property def version_info(self): """ @@ -265,8 +267,8 @@ def execute(self, command, with_keep_cwd=False, with_extended_output=False, with_exceptions=True, - as_process=False, - output_stream=None, + as_process=False, + output_stream=None, output_strip=True, **subprocess_kwargs ): @@ -301,7 +303,7 @@ def execute(self, command, wrapper that will interrupt the process once it goes out of scope. If you use the command in iterators, you should pass the whole process instance instead of a single stream. - + :param output_stream: If set to a file-like object, data produced by the git command will be output to the given stream directly. @@ -309,27 +311,27 @@ def execute(self, command, always be created with a pipe due to issues with subprocess. This merely is a workaround as data will be copied from the output pipe to the given output stream directly. - + :param output_strip: Strip the last line of the output if it is empty (default). Stripping should be disabled whenever it is important that the output is not modified in any way. For example when retrieving patch files using git-diff. - + :param subprocess_kwargs: Keyword arguments to be passed to subprocess.Popen. Please note that some of the valid kwargs are already set by this method, the ones you specify may not be the same ones. - + :return: * str(output) if extended_output = False (Default) * tuple(int(status), str(stdout), str(stderr)) if extended_output = True - + if ouput_stream is True, the stdout value will be your output stream: * output_stream if extended_output = False * tuple(int(status), output_stream, str(stderr)) if extended_output = True - + :raise GitCommandError: - + :note: If you add additional keyword arguments to the signature of this method, you must update the execute_kwargs tuple housed in this module.""" @@ -338,32 +340,32 @@ def execute(self, command, # Allow the user to have the command executed in their working dir. if with_keep_cwd or self._working_dir is None: - cwd = os.getcwd() + cwd = os.getcwd() else: - cwd=self._working_dir - + cwd = self._working_dir + # Start the process env = os.environ.copy() env['LANG'] = 'C' proc = Popen(command, - cwd=cwd, - stdin=istream, - stderr=PIPE, - stdout=PIPE, - close_fds=(os.name=='posix'),# unsupported on linux - env=env, - **subprocess_kwargs - ) + cwd=cwd, + stdin=istream, + stderr=PIPE, + stdout=PIPE, + close_fds=(os.name == 'posix'), # unsupported on linux + env=env, + **subprocess_kwargs + ) if as_process: return self.AutoInterrupt(proc, command) - + # Wait for the process to return status = 0 stdout_value = '' stderr_value = '' try: if output_stream is None: - stdout_value, stderr_value = proc.communicate() + stdout_value, stderr_value = proc.communicate() # strip trailing "\n" if stdout_value.endswith("\n") and output_strip: stdout_value = stdout_value[:-1] @@ -425,18 +427,18 @@ def transform_kwargs(self, **kwargs): @classmethod def __unpack_args(cls, arg_list): - if not isinstance(arg_list, (list,tuple)): + if not isinstance(arg_list, (list, tuple)): if isinstance(arg_list, unicode): return [arg_list.encode('utf-8')] - return [ str(arg_list) ] - + return [str(arg_list)] + outlist = list() for arg in arg_list: if isinstance(arg_list, (list, tuple)): - outlist.extend(cls.__unpack_args( arg )) + outlist.extend(cls.__unpack_args(arg)) elif isinstance(arg_list, unicode): outlist.append(arg_list.encode('utf-8')) - # END recursion + # END recursion else: outlist.append(str(arg)) # END for each arg @@ -475,16 +477,16 @@ def _call_process(self, method, *args, **kwargs): # Prepare the argument list opt_args = self.transform_kwargs(**kwargs) - + ext_args = self.__unpack_args([a for a in args if a is not None]) args = opt_args + ext_args - + def make_call(): call = [self.GIT_PYTHON_GIT_EXECUTABLE, dashify(method)] call.extend(args) return call - #END utility to recreate call after changes - + # END utility to recreate call after changes + if sys.platform == 'win32': try: try: @@ -493,33 +495,33 @@ def make_call(): # did we switch to git.cmd already, or was it changed from default ? permanently fail if self.GIT_PYTHON_GIT_EXECUTABLE != self.git_exec_name: raise - #END handle overridden variable + # END handle overridden variable type(self).GIT_PYTHON_GIT_EXECUTABLE = self.git_exec_name_win call = [self.GIT_PYTHON_GIT_EXECUTABLE] + list(args) - + try: return self.execute(make_call(), **_kwargs) finally: import warnings msg = "WARNING: Automatically switched to use git.cmd as git executable, which reduces performance by ~70%." - msg += "Its recommended to put git.exe into the PATH or to set the %s environment variable to the executable's location" % self._git_exec_env_var + msg += "Its recommended to put git.exe into the PATH or to set the %s environment variable to the executable's location" % self._git_exec_env_var warnings.warn(msg) - #END print of warning - #END catch first failure + # END print of warning + # END catch first failure except WindowsError: raise WindowsError("The system cannot find or execute the file at %r" % self.GIT_PYTHON_GIT_EXECUTABLE) - #END provide better error message + # END provide better error message else: return self.execute(make_call(), **_kwargs) - #END handle windows default installation - + # END handle windows default installation + def _parse_object_header(self, header_line): """ :param header_line: type_string size_as_int - + :return: (hex_sha, type_string, size_as_int) - + :raise ValueError: if the header contains indication for an error due to incorrect input sha""" tokens = header_line.split() @@ -530,46 +532,46 @@ def _parse_object_header(self, header_line): raise ValueError("SHA %s could not be resolved, git returned: %r" % (tokens[0], header_line.strip())) # END handle actual return value # END error handling - + if len(tokens[0]) != 40: - raise ValueError("Failed to parse header: %r" % header_line) + raise ValueError("Failed to parse header: %r" % header_line) return (tokens[0], tokens[1], int(tokens[2])) - + def __prepare_ref(self, ref): # required for command to separate refs on stdin refstr = str(ref) # could be ref-object if refstr.endswith("\n"): return refstr return refstr + "\n" - - def __get_persistent_cmd(self, attr_name, cmd_name, *args,**kwargs): + + def __get_persistent_cmd(self, attr_name, cmd_name, *args, **kwargs): cur_val = getattr(self, attr_name) if cur_val is not None: return cur_val - - options = { "istream" : PIPE, "as_process" : True } - options.update( kwargs ) - - cmd = self._call_process( cmd_name, *args, **options ) - setattr(self, attr_name, cmd ) + + options = {"istream": PIPE, "as_process": True} + options.update(kwargs) + + cmd = self._call_process(cmd_name, *args, **options) + setattr(self, attr_name, cmd) return cmd - + def __get_object_header(self, cmd, ref): cmd.stdin.write(self.__prepare_ref(ref)) cmd.stdin.flush() return self._parse_object_header(cmd.stdout.readline()) - + def get_object_header(self, ref): """ Use this method to quickly examine the type and size of the object behind the given ref. - + :note: The method will only suffer from the costs of command invocation once and reuses the command in subsequent calls. - + :return: (hexsha, type_string, size_as_int)""" cmd = self.__get_persistent_cmd("cat_file_header", "cat_file", batch_check=True) return self.__get_object_header(cmd, ref) - + def get_object_data(self, ref): """ As get_object_header, but returns object data as well :return: (hexsha, type_string, size_as_int,data_string) @@ -578,7 +580,7 @@ def get_object_data(self, ref): data = stream.read(size) del(stream) return (hexsha, typename, size, data) - + def stream_object_data(self, ref): """As get_object_header, but returns the data as a stream :return: (hexsha, type_string, size_as_int, stream) @@ -587,12 +589,12 @@ def stream_object_data(self, ref): cmd = self.__get_persistent_cmd("cat_file_all", "cat_file", batch=True) hexsha, typename, size = self.__get_object_header(cmd, ref) return (hexsha, typename, size, self.CatFileContentStream(size, cmd.stdout)) - + def clear_cache(self): """Clear all kinds of internal caches to release resources. - + Currently persistent commands will be interrupted. - + :return: self""" self.cat_file_all = None self.cat_file_header = None diff --git a/git/config.py b/git/config.py index e5cba9362..c66d0cc4d 100644 --- a/git/config.py +++ b/git/config.py @@ -17,7 +17,9 @@ __all__ = ('GitConfigParser', 'SectionConstraint') + class MetaParserBuilder(type): + """Utlity class wrapping base-class methods into decorators that assure read-only properties""" def __new__(metacls, name, bases, clsdict): """ @@ -27,7 +29,7 @@ def __new__(metacls, name, bases, clsdict): if kmm in clsdict: mutating_methods = clsdict[kmm] for base in bases: - methods = ( t for t in inspect.getmembers(base, inspect.ismethod) if not t[0].startswith("_") ) + methods = (t for t in inspect.getmembers(base, inspect.ismethod) if not t[0].startswith("_")) for name, method in methods: if name in clsdict: continue @@ -35,30 +37,32 @@ def __new__(metacls, name, bases, clsdict): if name in mutating_methods: method_with_values = set_dirty_and_flush_changes(method_with_values) # END mutating methods handling - + clsdict[name] = method_with_values # END for each name/method pair # END for each base # END if mutating methods configuration is set - + new_type = super(MetaParserBuilder, metacls).__new__(metacls, name, bases, clsdict) return new_type - - + def needs_values(func): """Returns method assuring we read values (on demand) before we try to access them""" + def assure_data_present(self, *args, **kwargs): self.read() return func(self, *args, **kwargs) # END wrapper method assure_data_present.__name__ = func.__name__ return assure_data_present - + + def set_dirty_and_flush_changes(non_const_func): """Return method that checks whether given non constant function may be called. If so, the instance will be set dirty. Additionally, we flush the changes right to disk""" + def flush_changes(self, *args, **kwargs): rval = non_const_func(self, *args, **kwargs) self.write() @@ -66,64 +70,65 @@ def flush_changes(self, *args, **kwargs): # END wrapper method flush_changes.__name__ = non_const_func.__name__ return flush_changes - + class SectionConstraint(object): + """Constrains a ConfigParser to only option commands which are constrained to always use the section we have been initialized with. - + It supports all ConfigParser methods that operate on an option""" __slots__ = ("_config", "_section_name") - _valid_attrs_ = ("get_value", "set_value", "get", "set", "getint", "getfloat", "getboolean", "has_option", - "remove_section", "remove_option", "options") - + _valid_attrs_ = ("get_value", "set_value", "get", "set", "getint", "getfloat", "getboolean", "has_option", + "remove_section", "remove_option", "options") + def __init__(self, config, section): self._config = config self._section_name = section - + def __getattr__(self, attr): if attr in self._valid_attrs_: return lambda *args, **kwargs: self._call_config(attr, *args, **kwargs) - return super(SectionConstraint,self).__getattribute__(attr) - + return super(SectionConstraint, self).__getattribute__(attr) + def _call_config(self, method, *args, **kwargs): """Call the configuration at the given method which must take a section name as first argument""" return getattr(self._config, method)(self._section_name, *args, **kwargs) - + @property def config(self): """return: Configparser instance we constrain""" return self._config - + class GitConfigParser(cp.RawConfigParser, object): + """Implements specifics required to read git style configuration files. - + This variation behaves much like the git.config command such that the configuration will be read on demand based on the filepath given during initialization. - + The changes will automatically be written once the instance goes out of scope, but can be triggered manually as well. - + The configuration file will be locked if you intend to change values preventing other instances to write concurrently. - + :note: The config is case-sensitive even when queried, hence section and option names must match perfectly.""" __metaclass__ = MetaParserBuilder - - + #{ Configuration # The lock type determines the type of lock to use in new configuration readers. # They must be compatible to the LockFile interface. # A suitable alternative would be the BlockingLockFile t_lock = LockFile re_comment = re.compile('^\s*[#;]') - - #} END configuration - + + #} END configuration + OPTCRE = re.compile( r'\s*(?P