Copyright (C) 2010 Google Inc.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. Service details and instances for the Docs service using GData 3.0.
Some use cases: Upload a document: docs upload --folder "Some folder" path_to_doc
Edit a document in your word editor: docs edit --title "Grocery List" --editor vim (editor also set in prefs)
Download docs: docs get --folder "Some folder"
from __future__ import with_statement
__author__ = 'tom.h.miller@gmail.com (Tom Miller)'
import gdata.docs.client
import logging
import os
import shutil
import googlecl
import googlecl.client
from googlecl.docs import SECTION_HEADER
import googlecl.docs.base
import atom.data
LOG = logging.getLogger(googlecl.docs.LOGGER_NAME + '.client')
Extends gdata.docs.client.DocsClient for the command line.
class DocsClientCL(gdata.docs.client.DocsClient,
googlecl.docs.base.DocsBaseCL,
googlecl.client.BaseClientCL):
This class adds some features focused on using Google Docs via an installed app with a command line interface.
DOCLIST_FEED_URI = '/feeds/default/private/full'
File extension/mimetype pairs of common format. These seem to have disappeared in python-gdata 2.0.15 and 2.0.16, so here they are given explicitly.
MIMETYPES = {
'CSV': 'text/csv',
'TSV': 'text/tab-separated-values',
'TAB': 'text/tab-separated-values',
'DOC': 'application/msword',
'DOCX': ('application/vnd.openxmlformats-officedocument.'
'wordprocessingml.document'),
'ODS': 'application/x-vnd.oasis.opendocument.spreadsheet',
'ODT': 'application/vnd.oasis.opendocument.text',
'RTF': 'application/rtf',
'SXW': 'application/vnd.sun.xml.writer',
'TXT': 'text/plain',
'XLS': 'application/vnd.ms-excel',
'XLSX': 'application/vnd.openxmlformats-officedocument.spreadsheetml.sheet',
'PDF': 'application/pdf',
'PNG': 'image/png',
'PPT': 'application/vnd.ms-powerpoint',
'PPS': 'application/vnd.ms-powerpoint',
'HTM': 'text/html',
'HTML': 'text/html',
'ZIP': 'application/zip',
'SWF': 'application/x-shockwave-flash'
}
Constructor.
def __init__(self, config):
gdata.docs.client.DocsClient.__init__(self, source='GoogleCL')
googlecl.client.BaseClientCL.__init__(self, SECTION_HEADER, config)
Wrapper function to mesh with DocsBaseCL.upload_docs().
def _create_folder(self, title, folder_or_uri):
return self.create(gdata.docs.data.FOLDER_LABEL, title,
folder_or_uri)
def _determine_content_type(self, file_ext):
try:
return DocsClientCL.MIMETYPES[file_ext.upper()]
except KeyError:
LOG.info('No supported filetype found for extension %s', file_ext)
return None
Downloads a file, optionally decoding from UTF-8.
def _download_file(self, uri, file_path, auth_token=None, **kwargs):
Overridden from gdata.docs.client to support decoding.
Args: uri: string The full Export URL to download the file from. file_path: string The full path to save the file to. auth_token: (optional) gdata.gauth.ClientLoginToken, AuthSubToken, or OAuthToken which authorizes this client to edit the user's data. decode: bool (default False) Whether or not to decode UTF-8. kwargs: Other parameters to pass to self.get_file_content().
Raises: RequestError: on error response from server.
response_string = self.get_file_content(uri, auth_token=auth_token)
if googlecl.docs.base.can_export(uri) and\
self.config.lazy_get(SECTION_HEADER, 'decode_utf_8', False, bool):
try:
file_string = response_string.decode('utf-8-sig')
except UnicodeError, err:
LOG.debug('Could not decode: ' + str(err))
file_string = response_string
else:
file_string = response_string
with open(file_path, 'wb') as download_file:
download_file.write(file_string)
download_file.flush()
Exports a document from the Document List in a different format.
def export(self, entry_or_id_or_url, file_path, gid=None, auth_token=None,
**kwargs):
Overloaded from docs.client.DocsClient to fix "always-download-as-pdf" issue
Args: entry_or_id_or_url: gdata.data.GDEntry or string representing a resource id or URL to download the document from (such as the content src link). file_path: str The full path to save the file to. The export format is inferred from the the file extension. gid: str (optional) grid id for downloading a single grid of a spreadsheet. The param should only be used for .csv and .tsv spreadsheet exports. auth_token: (optional) gdata.gauth.ClientLoginToken, AuthSubToken, or OAuthToken which authorizes this client to edit the user's data. kwargs: Other parameters to pass to self.download().
Raises: gdata.client.RequestError if the download URL is malformed or the server's response was not successful.
extra_params = {}
match = gdata.docs.data.FILE_EXT_PATTERN.match(file_path)
if match:
export_format = match.group(1)
Hack for apps-api-issues Issue 2294
if export_format.lower() == 'html':
export_format = '102'
extra_params['exportFormat'] = export_format
Fix issue with new-style docs always downloading to PDF (gdata-issues Issue 2157)
extra_params['format'] = export_format
if gid is not None:
extra_params['gid'] = gid
self.download(entry_or_id_or_url, file_path, extra_params,
auth_token=auth_token, **kwargs)
Export = export
Get a list of document entries from a feed.
def get_doclist(self, titles=None, folder_entry_list=None):
Keyword arguments: titles: list or string Title(s) of entries to return. Will be compared to entry.title.text, using regular expressions if self.use_regex. Default None for all entries from feed. folder_entry_list: List of GDataEntry's of folders to get from. Only files found in these folders will be returned. Default None for all folders.
Returns: List of entries.
if folder_entry_list:
entries = []
for folder in folder_entry_list:
folder.content.src is the uri to query for documents in that folder.
entries.extend(self.GetEntries(folder.content.src,
titles,
desired_class=gdata.docs.data.DocList))
else:
entries = self.GetEntries(gdata.docs.client.DOCLIST_FEED_URI,
titles,
desired_class=gdata.docs.data.DocList)
return entries
Return exactly one doc_entry.
def get_single_doc(self, title=None, folder_entry_list=None):
Keyword arguments: title: Title to match on for document. Default None for any title. folder_entry_list: GDataEntry of folders to look in. Default None for any folder.
Returns: None if there were no matches, or one entry matching the given title.
if folder_entry_list:
if len(folder_entry_list) == 1:
return self.GetSingleEntry(folder_entry_list[0].content.src,
title,
desired_class=gdata.docs.data.DocList)
else:
entries = self.get_doclist(title, folder_entry_list)
Technically don't need the desired_class for this call because we have the entries.
return self.GetSingleEntry(entries, title)
else:
return self.GetSingleEntry(gdata.docs.client.DOCLIST_FEED_URI,
title,
desired_class=gdata.docs.data.DocList)
GetSingleDoc = get_single_doc
Return entries for one or more folders.
def get_folder(self, title):
Keyword arguments: title: Title of the folder.
Returns: GDataEntry representing a folder, or None of title is None.
if title:
uri = gdata.docs.client.DOCLIST_FEED_URI + '-/folder'
folder_entries = self.GetEntries(uri, title)
if not folder_entries:
LOG.warning('No folder found that matches ' + title)
return folder_entries
else:
return None
GetFolder = get_folder
Check that the token being used is valid.
def is_token_valid(self, test_uri=None):
if not test_uri:
docs_uri = gdata.docs.client.DOCLIST_FEED_URI
sheets_uri = ('https://spreadsheets.google.com/feeds/spreadsheets'
'/private/full')
docs_test = googlecl.client.BaseClientCL.IsTokenValid(self, docs_uri)
sheets_test = googlecl.client.BaseClientCL.IsTokenValid(self, sheets_uri)
return docs_test and sheets_test
IsTokenValid = is_token_valid
Replace content of a DocEntry.
def _modify_entry(self, doc_entry, path_to_new_content, file_ext):
Args: doc_entry: DocEntry whose content will be replaced. path_to_new_content: str Path to file that has new content. file_ext: str Extension to use to determine MIME type of upload (e.g. 'txt', 'doc')
from gdata.docs.data import MIMETYPES
try:
content_type = MIMETYPES[file_ext.upper()]
except KeyError:
print 'Could not find mimetype for ' + file_ext
while file_ext not in MIMETYPES.keys():
file_ext = raw_input('Please enter one of ' +
MIMETYPES.keys() +
' to determine the content type to upload as.')
content_type = MIMETYPES[file_ext.upper()]
mediasource = gdata.data.MediaSource(file_path=path_to_new_content,
content_type=content_type)
return self.Update(doc_entry, media_source=mediasource)
Request access as in BaseClientCL, but specify scopes.
def request_access(self, domain, display_name, scopes=None, browser=None):
When people use docs (writely), they expect access to spreadsheets as well (wise).
if not scopes:
scopes = gdata.gauth.AUTH_SCOPES['writely'] +\
gdata.gauth.AUTH_SCOPES['wise']
return googlecl.client.BaseClientCL.request_access(self, domain,
display_name,
scopes=scopes,
browser=browser)
RequestAccess = request_access
Upload a document.
def _transmit_doc(self, path, entry_title, post_uri, content_type, file_ext):
The final step in uploading a document. The process differs between versions of the gdata python client library, hence its definition here.
Args: path: Path to the file to upload. entry_title: Name of the document. post_uri: URI to make request to. content_type: MIME type of request. file_ext: File extension that determined the content_type.
Returns: Entry representing the document uploaded.
GoogleCL that uses gdata-2.0.0 through 2.0.4 won't ever see this code. If it uses gdata-2.0.5 through 2.0.7, it would otherwise give an error about a resumable uploader that it doesn't have. This avoids that error. If it uses gdata-2.0.8, 2.0.9, or 2.0.11 it can't upload docs due to an SSL error. If it uses gdata-2.0.10, 2.0.12, 2.0.13, 2.0.14 this should allow it to upload all allowable file types.
if hasattr(gdata.client,"ResumableUploader"):
f = open(path)
file_size = os.path.getsize(f.name)
uploader = gdata.client.ResumableUploader(
self, f, content_type, file_size, chunk_size=1048576,
desired_class=gdata.data.GDEntry)
Set metadata for our upload.
entry = gdata.data.GDEntry(title=atom.data.Title(text=entry_title))
new_entry = uploader.UploadFile('/feeds/upload/create-session/default/private/full', entry=entry)
These might be useful for a verbose debug statement: print 'Document uploaded: ' + new_entry.title.text print 'Quota used: %s' % new_entry.quota_bytes_used.text
f.close()
return new_entry
else:
If we have reached this point, we must be in gdata-2.0.5 through 2.0.7 The upload is guaranteed to fail, so the self.upload call is here to return whatever the caller wanted.
return self.upload(path, entry_title, post_uri, content_type)
SERVICE_CLASS = DocsClientCL