Merge branch 'master' into dj111
commit
67652c2f72
|
@ -7,3 +7,11 @@ https://www.larcommons.net/sitemap.xml
|
|||
https://www.uwestminsterpress.co.uk/sitemap.xml
|
||||
https://www.stockholmuniversitypress.se/sitemap.xml
|
||||
https://www.luminosoa.org/sitemap.xml
|
||||
https://iitikship.iiti.ac.in/sitemap.xml
|
||||
https://aperio.press/sitemap.xml
|
||||
https://press.lse.ac.uk/sitemap.xml
|
||||
https://press.sjms.nu/sitemap.xml
|
||||
https://trystingtree.library.oregonstate.edu/sitemap.xml
|
||||
https://publishing.vt.edu/sitemap.xml
|
||||
https://universitypress.whiterose.ac.uk/sitemap.xml
|
||||
https://www.winchesteruniversitypress.org/sitemap.xml
|
|
@ -909,10 +909,9 @@ class BasePandataLoader(object):
|
|||
if work and id.work and id.work_id is not work.id:
|
||||
# dangerous! merge newer into older
|
||||
if work.id < id.work_id:
|
||||
merge_works(work, id.work)
|
||||
work = merge_works(work, id.work)
|
||||
else:
|
||||
merge_works(id.work, work)
|
||||
work = id.work
|
||||
work = merge_works(id.work, work)
|
||||
else:
|
||||
work = id.work
|
||||
if id.edition and not edition:
|
||||
|
@ -986,6 +985,7 @@ class BasePandataLoader(object):
|
|||
|
||||
def load_ebooks(self, metadata, edition, test_mode=False, user=None):
|
||||
default_edition = edition
|
||||
license = cc.license_from_cc_url(metadata.rights_url)
|
||||
for key in ['epub', 'pdf', 'mobi']:
|
||||
url = metadata.metadata.get('download_url_{}'.format(key), None)
|
||||
if url:
|
||||
|
@ -995,7 +995,6 @@ class BasePandataLoader(object):
|
|||
if contentfile:
|
||||
contentfile_name = '/loaded/ebook_{}.{}'.format(edition.id, key)
|
||||
path = default_storage.save(contentfile_name, contentfile)
|
||||
license = cc.license_from_cc_url(metadata.rights_url)
|
||||
ebf = models.EbookFile.objects.create(
|
||||
format=key,
|
||||
edition=edition,
|
||||
|
|
|
@ -65,7 +65,7 @@
|
|||
"goodreads_auth_token": null,
|
||||
"goodreads_user_link": null,
|
||||
"user": 1,
|
||||
"facebook_id": null,
|
||||
"facebook_id": "",
|
||||
"librarything_id": "",
|
||||
"home_url": "",
|
||||
"pic_url": "",
|
||||
|
@ -85,7 +85,7 @@
|
|||
"goodreads_auth_token": null,
|
||||
"goodreads_user_link": null,
|
||||
"user": 2,
|
||||
"facebook_id": null,
|
||||
"facebook_id": "",
|
||||
"librarything_id": "",
|
||||
"home_url": "",
|
||||
"pic_url": "",
|
||||
|
@ -105,7 +105,7 @@
|
|||
"goodreads_auth_token": null,
|
||||
"goodreads_user_link": null,
|
||||
"user": 3,
|
||||
"facebook_id": null,
|
||||
"facebook_id": "",
|
||||
"librarything_id": "",
|
||||
"home_url": "",
|
||||
"pic_url": "",
|
||||
|
|
|
@ -65,7 +65,12 @@ def store_doab_cover(doab_id, redo=False):
|
|||
else:
|
||||
r = requests.get(url)
|
||||
cover_file = ContentFile(r.content)
|
||||
cover_file.content_type = r.headers.get('content-type', '')
|
||||
content_type = r.headers.get('content-type', '')
|
||||
if u'text/html' in content_type:
|
||||
logger.warning('Cover return html for doab_id={}: {}'.format(doab_id, e))
|
||||
return (None, False)
|
||||
cover_file.content_type = content_type
|
||||
|
||||
|
||||
default_storage.save(cover_file_name, cover_file)
|
||||
return (default_storage.url(cover_file_name), True)
|
||||
|
@ -287,6 +292,12 @@ def load_doab_edition(title, doab_id, url, format, rights,
|
|||
publisher_name=unlist(kwargs.get('publisher')),
|
||||
authors=kwargs.get('creator'),
|
||||
)
|
||||
if rights:
|
||||
for ebook in edition.ebooks.all():
|
||||
if not ebook.rights:
|
||||
ebook.rights = rights
|
||||
ebook.save()
|
||||
|
||||
return edition
|
||||
|
||||
#
|
||||
|
|
|
@ -124,6 +124,9 @@ def online_to_download(url):
|
|||
booknum = FRONTIERSIN.search(url).group(1)
|
||||
urls.append(u'https://www.frontiersin.org/GetFile.aspx?ebook={}&fileformat=EPUB'.format(booknum))
|
||||
urls.append(u'https://www.frontiersin.org/GetFile.aspx?ebook={}&fileformat=PDF'.format(booknum))
|
||||
elif url.find(u'edp-open.org/books-in-') >= 0:
|
||||
# pages needing multi-scrape
|
||||
return urls
|
||||
else:
|
||||
urls.append(url)
|
||||
return urls
|
||||
|
|
|
@ -0,0 +1,95 @@
|
|||
import logging
|
||||
import re
|
||||
from urlparse import urljoin
|
||||
|
||||
from bs4 import BeautifulSoup
|
||||
import requests
|
||||
|
||||
from django.conf import settings
|
||||
|
||||
from regluit.core.bookloader import add_from_bookdatas
|
||||
from regluit.core.loaders.scrape import BaseScraper
|
||||
from regluit.core.validation import identifier_cleaner
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
'''
|
||||
use for web pages with multiple books
|
||||
returns an iterator of scrapers
|
||||
'''
|
||||
|
||||
class BaseMultiScraper(BaseScraper):
|
||||
def __init__(self, url, doc):
|
||||
self.metadata = {}
|
||||
self.identifiers = {'http': url}
|
||||
self.doc = doc
|
||||
self.base = url
|
||||
self.get_all()
|
||||
if not self.metadata.get('title', None):
|
||||
self.set('title', '!!! missing title !!!')
|
||||
self.metadata['identifiers'] = self.identifiers
|
||||
|
||||
def multiscrape(url, divider, scraper_class=BaseMultiScraper):
|
||||
try:
|
||||
response = requests.get(url, headers={"User-Agent": settings.USER_AGENT})
|
||||
if response.status_code == 200:
|
||||
doc = BeautifulSoup(response.content, 'lxml')
|
||||
sections = divider(doc)
|
||||
for section in sections:
|
||||
yield scraper_class(url, section)
|
||||
except requests.exceptions.RequestException as e:
|
||||
logger.error(e)
|
||||
self.metadata = None
|
||||
|
||||
|
||||
# following is code specific to edp-open.org; refactor when we add another
|
||||
|
||||
def divider(doc):
|
||||
return doc.select('article.Bk')
|
||||
|
||||
ISBNMATCH = re.compile(r'([\d\-]+)')
|
||||
class EDPMultiScraper(BaseMultiScraper):
|
||||
def get_isbns(self):
|
||||
'''return a dict of edition keys and ISBNs'''
|
||||
isbns = {}
|
||||
isbn_cleaner = identifier_cleaner('isbn', quiet=True)
|
||||
labels = ['epub', 'pdf', 'paper']
|
||||
info = self.doc.select_one('p.nfo').text
|
||||
isbntexts = re.split('ISBN', info)
|
||||
for isbntext in isbntexts[1:]:
|
||||
isbnmatch = ISBNMATCH.search(isbntext)
|
||||
if isbnmatch:
|
||||
isbn = isbn_cleaner(isbnmatch.group(0))
|
||||
isbns[labels.pop()] = isbn
|
||||
return isbns
|
||||
|
||||
def get_downloads(self):
|
||||
dl = self.doc.select_one('nav.dl')
|
||||
links = dl.select('a.fulldl')
|
||||
for link in links:
|
||||
href = urljoin(self.base, link['href'])
|
||||
if href.endswith('.pdf'):
|
||||
self.set('download_url_pdf', href)
|
||||
elif href.endswith('.epub'):
|
||||
self.set('download_url_epub', href)
|
||||
|
||||
def get_language(self):
|
||||
if 'english' in self.base:
|
||||
self.set('language', 'en')
|
||||
else:
|
||||
self.set('language', 'fr')
|
||||
|
||||
def get_title(self):
|
||||
value = self.doc.select_one('h2').text
|
||||
book_id = self.doc.select_one('h2')['id']
|
||||
self.identifiers['http'] = u'{}#{}'.format(self.base, book_id)
|
||||
self.set('title', value)
|
||||
|
||||
def edp_scrape():
|
||||
edp_urls = [
|
||||
'https://www.edp-open.org/books-in-french',
|
||||
'https://www.edp-open.org/books-in-english',
|
||||
]
|
||||
for url in edp_urls:
|
||||
scrapers = multiscrape(url, divider, scraper_class=EDPMultiScraper)
|
||||
add_from_bookdatas(scrapers)
|
||||
|
|
@ -51,19 +51,7 @@ class BaseScraper(object):
|
|||
self.doc = BeautifulSoup(response.content, 'lxml')
|
||||
for review in self.doc.find_all(itemtype="http://schema.org/Review"):
|
||||
review.clear()
|
||||
self.setup()
|
||||
self.get_genre()
|
||||
self.get_title()
|
||||
self.get_language()
|
||||
self.get_description()
|
||||
self.get_identifiers()
|
||||
self.get_keywords()
|
||||
self.get_publisher()
|
||||
self.get_pubdate()
|
||||
self.get_authors()
|
||||
self.get_cover()
|
||||
self.get_downloads()
|
||||
self.get_license()
|
||||
self.get_all()
|
||||
if not self.metadata.get('title', None):
|
||||
self.set('title', '!!! missing title !!!')
|
||||
if not self.metadata.get('language', None):
|
||||
|
@ -141,6 +129,21 @@ class BaseScraper(object):
|
|||
value_list.append(el['content'])
|
||||
return value_list
|
||||
|
||||
def get_all(self):
|
||||
self.setup()
|
||||
self.get_genre()
|
||||
self.get_title()
|
||||
self.get_language()
|
||||
self.get_description()
|
||||
self.get_identifiers()
|
||||
self.get_keywords()
|
||||
self.get_publisher()
|
||||
self.get_pubdate()
|
||||
self.get_authors()
|
||||
self.get_cover()
|
||||
self.get_downloads()
|
||||
self.get_license()
|
||||
|
||||
def setup(self):
|
||||
# use this method to get auxiliary resources based on doc
|
||||
pass
|
||||
|
|
|
@ -8,7 +8,9 @@ from . import BaseScraper
|
|||
HAS_EDS = re.compile(r'\(eds?\.\)')
|
||||
UBIQUITY_HOSTS = ["ubiquitypress.com", "kriterium.se", "oa.finlit.fi", "humanities-map.net",
|
||||
"oa.psupress.org", "larcommons.net", "uwestminsterpress.co.uk", "stockholmuniversitypress.se",
|
||||
"luminosoa.org",
|
||||
"luminosoa.org", "iitikship.iiti.ac.in", "aperio.press", "press.lse.ac.uk", "press.sjms.nu",
|
||||
"trystingtree.library.oregonstate.edu", "publishing.vt.edu", "universitypress.whiterose.ac.uk",
|
||||
"www.winchesteruniversitypress.org",
|
||||
]
|
||||
|
||||
class UbiquityScraper(BaseScraper):
|
||||
|
|
|
@ -0,0 +1,22 @@
|
|||
import string
|
||||
from django.core.management.base import BaseCommand
|
||||
from regluit.core.models import FACEBOOK, UNGLUEITAR
|
||||
from regluit.libraryauth.auth import pic_storage_url
|
||||
|
||||
from regluit.core import models
|
||||
|
||||
class Command(BaseCommand):
|
||||
help = "fix avatar urls and settings"
|
||||
|
||||
def handle(self, **options):
|
||||
for profile in models.UserProfile.objects.exclude(pic_url=''):
|
||||
print "updating user %s" % profile.user
|
||||
if not profile.pic_url.startswith('https://unglueit'):
|
||||
profile.pic_url = pic_storage_url(profile.user, 'twitter', profile.pic_url)
|
||||
profile.save()
|
||||
for profile in models.UserProfile.objects.filter(avatar_source=FACEBOOK):
|
||||
print "updating user %s" % profile.user
|
||||
profile.facebook_id = ''
|
||||
if not profile.pic_url:
|
||||
profile.avatar_source = UNGLUEITAR
|
||||
profile.save()
|
|
@ -1,16 +0,0 @@
|
|||
import string
|
||||
from django.core.management.base import BaseCommand
|
||||
from regluit.core.models import TWITTER
|
||||
|
||||
from regluit.core import models
|
||||
|
||||
class Command(BaseCommand):
|
||||
help = "fix old twitter avatar urls"
|
||||
|
||||
def handle(self, **options):
|
||||
print "Number of users affected with : %s" % models.UserProfile.objects.filter( pic_url__contains='//si0.twimg.com').count()
|
||||
|
||||
for profile in models.UserProfile.objects.filter(pic_url__contains='//si0.twimg.com'):
|
||||
print "updating user %s" % profile.user
|
||||
profile.pic_url = string.replace( profile.pic_url, '//si0.twimg.com','//pbs.twimg.com')
|
||||
profile.save()
|
|
@ -0,0 +1,10 @@
|
|||
from django.core.management.base import BaseCommand
|
||||
|
||||
from regluit.core.loaders.multiscrape import edp_scrape
|
||||
|
||||
|
||||
class Command(BaseCommand):
|
||||
help = "load books from edp-open"
|
||||
|
||||
def handle(self, **options):
|
||||
edp_scrape()
|
|
@ -0,0 +1,25 @@
|
|||
# -*- coding: utf-8 -*-
|
||||
from __future__ import unicode_literals
|
||||
|
||||
from django.db import migrations, models
|
||||
|
||||
|
||||
class Migration(migrations.Migration):
|
||||
|
||||
dependencies = [
|
||||
('core', '0014_auto_20180618_1646'),
|
||||
]
|
||||
|
||||
operations = [
|
||||
migrations.AlterField(
|
||||
model_name='userprofile',
|
||||
name='avatar_source',
|
||||
field=models.PositiveSmallIntegerField(default=4, null=True, choices=[(0, b'No Avatar, Please'), (1, b'Gravatar'), (2, b'Twitter/Facebook'), (4, b'Unglueitar')]),
|
||||
),
|
||||
migrations.AlterField(
|
||||
model_name='userprofile',
|
||||
name='facebook_id',
|
||||
field=models.CharField(default='', max_length=31, blank=True),
|
||||
preserve_default=False,
|
||||
),
|
||||
]
|
|
@ -1163,7 +1163,7 @@ class UserProfile(models.Model):
|
|||
pic_url = models.URLField(blank=True)
|
||||
home_url = models.URLField(blank=True)
|
||||
twitter_id = models.CharField(max_length=15, blank=True)
|
||||
facebook_id = models.BigIntegerField(null=True, blank=True)
|
||||
facebook_id = models.CharField(max_length=31, blank=True)
|
||||
librarything_id = models.CharField(max_length=31, blank=True)
|
||||
badges = models.ManyToManyField('Badge', related_name='holders', blank=True)
|
||||
kindle_email = models.EmailField(max_length=254, blank=True)
|
||||
|
@ -1183,8 +1183,7 @@ class UserProfile(models.Model):
|
|||
choices=(
|
||||
(NO_AVATAR, 'No Avatar, Please'),
|
||||
(GRAVATAR, 'Gravatar'),
|
||||
(TWITTER, 'Twitter'),
|
||||
(FACEBOOK, 'Facebook'),
|
||||
(TWITTER, 'Twitter/Facebook'),
|
||||
(UNGLUEITAR, 'Unglueitar'),
|
||||
)
|
||||
)
|
||||
|
@ -1314,7 +1313,7 @@ class UserProfile(models.Model):
|
|||
|
||||
@property
|
||||
def avatar_url(self):
|
||||
if self.avatar_source is None or self.avatar_source is TWITTER:
|
||||
if self.avatar_source is None or self.avatar_source in (TWITTER, FACEBOOK):
|
||||
if self.pic_url:
|
||||
return self.pic_url
|
||||
else:
|
||||
|
@ -1323,9 +1322,6 @@ class UserProfile(models.Model):
|
|||
return self.unglueitar()
|
||||
elif self.avatar_source == GRAVATAR:
|
||||
return self.gravatar()
|
||||
elif self.avatar_source == FACEBOOK and self.facebook_id != None:
|
||||
return 'https://graph.facebook.com/v2.3/' + str(self.facebook_id) + '/picture?redirect=true'
|
||||
else:
|
||||
return ANONYMOUS_AVATAR
|
||||
|
||||
@property
|
||||
|
|
|
@ -866,39 +866,47 @@ class Edition(models.Model):
|
|||
def cover_image_large(self):
|
||||
#550 pixel high image
|
||||
if self.cover_image:
|
||||
try:
|
||||
im = get_thumbnail(self.cover_image, 'x550', crop='noop', quality=95)
|
||||
if im.exists():
|
||||
return im.url
|
||||
except IOError:
|
||||
pass
|
||||
elif self.googlebooks_id:
|
||||
url = "https://encrypted.google.com/books?id=%s&printsec=frontcover&img=1&zoom=0" % self.googlebooks_id
|
||||
try:
|
||||
im = get_thumbnail(url, 'x550', crop='noop', quality=95)
|
||||
if not im.exists() or im.storage.size(im.name) == 16392: # check for "image not available" image
|
||||
url = "https://encrypted.google.com/books?id=%s&printsec=frontcover&img=1&zoom=1" % self.googlebooks_id
|
||||
im = get_thumbnail(url, 'x550', crop='noop', quality=95)
|
||||
if im.exists():
|
||||
return im.url
|
||||
else:
|
||||
return ''
|
||||
else:
|
||||
except IOError:
|
||||
pass
|
||||
return ''
|
||||
|
||||
def cover_image_small(self):
|
||||
#80 pixel high image
|
||||
if self.cover_image:
|
||||
try:
|
||||
im = get_thumbnail(self.cover_image, 'x80', crop='noop', quality=95)
|
||||
if im.exists():
|
||||
return im.url
|
||||
except IOError:
|
||||
pass
|
||||
if self.googlebooks_id:
|
||||
return "https://encrypted.google.com/books?id=%s&printsec=frontcover&img=1&zoom=5" % self.googlebooks_id
|
||||
else:
|
||||
return ''
|
||||
|
||||
def cover_image_thumbnail(self):
|
||||
#128 pixel wide image
|
||||
if self.cover_image:
|
||||
try:
|
||||
im = get_thumbnail(self.cover_image, '128', crop='noop', quality=95)
|
||||
if im.exists():
|
||||
return im.url
|
||||
except IOError:
|
||||
pass
|
||||
if self.googlebooks_id:
|
||||
return "https://encrypted.google.com/books?id=%s&printsec=frontcover&img=1&zoom=1" % self.googlebooks_id
|
||||
else:
|
||||
|
|
|
@ -188,7 +188,7 @@ class ProfileForm(forms.ModelForm):
|
|||
|
||||
class Meta:
|
||||
model = UserProfile
|
||||
fields = 'tagline', 'librarything_id', 'home_url', 'clear_facebook', 'clear_twitter', 'clear_goodreads', 'avatar_source'
|
||||
fields = 'tagline', 'librarything_id', 'facebook_id', 'home_url', 'clear_facebook', 'clear_twitter', 'clear_goodreads', 'avatar_source'
|
||||
widgets = {
|
||||
'tagline': forms.Textarea(attrs={'rows': 5, 'onKeyUp': "counter(this, 140)", 'onBlur': "counter(this, 140)"}),
|
||||
}
|
||||
|
@ -198,22 +198,12 @@ class ProfileForm(forms.ModelForm):
|
|||
super(ProfileForm, self).__init__(*args, **kwargs)
|
||||
choices = []
|
||||
for choice in self.fields['avatar_source'].choices :
|
||||
if choice[0] == FACEBOOK and not profile.facebook_id:
|
||||
pass
|
||||
elif choice[0] == TWITTER and not profile.twitter_id:
|
||||
if choice[0] == TWITTER and not profile.pic_url:
|
||||
pass
|
||||
else:
|
||||
choices.append(choice)
|
||||
self.fields['avatar_source'].choices = choices
|
||||
|
||||
def clean(self):
|
||||
# check that if a social net is cleared, we're not using it a avatar source
|
||||
if self.cleaned_data.get("clear_facebook", False) and self.cleaned_data.get("avatar_source", None) == FACEBOOK:
|
||||
self.cleaned_data["avatar_source"] == UNGLUEITAR
|
||||
if self.cleaned_data.get("clear_twitter", False) and self.cleaned_data.get("avatar_source", None) == TWITTER:
|
||||
self.cleaned_data["avatar_source"] == UNGLUEITAR
|
||||
return self.cleaned_data
|
||||
|
||||
def getTransferCreditForm(maximum, data=None, *args, **kwargs ):
|
||||
class TransferCreditForm(forms.Form):
|
||||
recipient = AutoCompleteSelectField(
|
||||
|
|
|
@ -86,7 +86,7 @@ We use <a href="https://stripe.com/">Stripe</a> to collect payments. As a result
|
|||
</li>
|
||||
|
||||
<li>
|
||||
We use avatar images from <a href="https://twitter.com/">Twitter</a>, <a href="https://facebook.com/">Facebook</a>, and <a href="https://gravatar.com/">Automattic's Gravatar service</a>. If you see an avatar on a page at unglue.it and you use an older web browser that doesn't use Referrer meta tags, one or more of these companies can tell what page on our site you're looking at. As you're probably aware, facebook <a href="https://www.facebook.com/policy.php">doesn't put much stock</a> in privacy. You can judge privacy policies at <a href="https://twitter.com/privacy?lang=en">Twitter</a> and <a href="http://automattic.com/privacy/">Automattic</a> for yourself.
|
||||
We use avatar images from <a href="https://gravatar.com/">Automattic's Gravatar service</a>. If you see an avatar on a page at unglue.it and you use an older web browser that doesn't use Referrer meta tags, Automattic can tell what page on our site you're looking at. You can judge the privacy policy at <a href="http://automattic.com/privacy/">Automattic</a> for yourself. We no longer use images from Facebook or Twitter services.
|
||||
</li>
|
||||
|
||||
<li>
|
||||
|
|
|
@ -185,17 +185,19 @@ function highlightTarget(targetdiv) {
|
|||
</div>
|
||||
<div class="check-list">
|
||||
{% if supporter.profile.twitter_id %}
|
||||
<a href="{% url 'social:begin' backend='twitter' %}?next=/">Update your Twitter connection</a> <br /> or disconnect Twitter: {{ profile_form.clear_twitter }}
|
||||
<a href="{% url 'social:begin' backend='twitter' %}?next={% url 'supporter' supporter %}#edit">Update your Twitter connection</a> <br /> or disconnect Twitter: {{ profile_form.clear_twitter }}
|
||||
{% else %}
|
||||
<a href="{% url 'social:begin' backend='twitter' %}?next=/">Connect your Twitter account</a> to Unglue.it
|
||||
<a href="{% url 'social:begin' backend='twitter' %}?next={% url 'supporter' supporter %}#edit">Connect your Twitter account</a> to Unglue.it
|
||||
{% endif %}
|
||||
</div>
|
||||
<div class="check-list">
|
||||
{% if supporter.profile.facebook_id %}
|
||||
<a href="{% url 'social:begin' backend='facebook' %}?next=/">Update your Facebook connection</a> <br /> or disconnect Facebook: {{ profile_form.clear_facebook }}
|
||||
<a href="{% url 'social:begin' backend='facebook' %}?next={% url 'supporter' supporter %}#edit">Update your Facebook connection</a> <br /> or disconnect Facebook: {{ profile_form.clear_facebook }}
|
||||
{% else %}
|
||||
<a href="{% url 'social:begin' backend='facebook' %}?next=/">Connect your Facebook account</a> to Unglue.it
|
||||
<a href="{% url 'social:begin' backend='facebook' %}?next={% url 'supporter' supporter %}#edit">Connect your Facebook account</a> to Unglue.it
|
||||
{% endif %}
|
||||
<label>{% if supporter.profile.facebook_id %}Change{% else %}Add{% endif %} your <a href='https://www.facebook.com/help/211813265517027'>Facebook Username</a>:</label>
|
||||
{{ profile_form.facebook_id }}{{ profile_form.facebook_id.errors }}
|
||||
</div>
|
||||
<div class="check-list" id="connectgr">
|
||||
{% if user.profile.goodreads_user_id %}
|
||||
|
|
|
@ -2236,7 +2236,7 @@ def goodreads_cb(request):
|
|||
profile.save() # is this needed?
|
||||
|
||||
# redirect to the Goodreads display page -- should observe some next later
|
||||
return HttpResponseRedirect(reverse('home'))
|
||||
return HttpResponseRedirect(reverse('supporter', args=[request.user]))
|
||||
|
||||
@require_POST
|
||||
@login_required
|
||||
|
|
|
@ -1,20 +1,41 @@
|
|||
import logging
|
||||
|
||||
import requests
|
||||
from django.http import HttpResponse
|
||||
from django.shortcuts import redirect
|
||||
from django.utils.http import urlquote
|
||||
|
||||
from social.pipeline.social_auth import associate_by_email
|
||||
from social.apps.django_app.default.models import UserSocialAuth
|
||||
from social.apps.django_app.middleware import SocialAuthExceptionMiddleware
|
||||
from social.exceptions import (AuthAlreadyAssociated,SocialAuthBaseException)
|
||||
from social.utils import social_logger
|
||||
from django.core.files.base import ContentFile
|
||||
from django.core.files.storage import default_storage
|
||||
from social_core.pipeline.social_auth import associate_by_email
|
||||
from social_django.models import UserSocialAuth
|
||||
from social_django.middleware import SocialAuthExceptionMiddleware
|
||||
from social_core.exceptions import (AuthAlreadyAssociated, SocialAuthBaseException)
|
||||
|
||||
ANONYMOUS_AVATAR = '/static/images/header/avatar.png'
|
||||
(NO_AVATAR, GRAVATAR, TWITTER, FACEBOOK, PRIVATETAR) = (0, 1, 2, 3, 4)
|
||||
AVATARS = (NO_AVATAR, GRAVATAR, TWITTER, FACEBOOK, PRIVATETAR)
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
def pic_storage_url(user, backend, url):
|
||||
pic_file_name = '/pic/{}/{}'.format(backend, user)
|
||||
# download cover image to cover_file
|
||||
try:
|
||||
r = requests.get(url)
|
||||
pic_file = ContentFile(r.content)
|
||||
content_type = r.headers.get('content-type', '')
|
||||
if u'text' in content_type:
|
||||
logger.warning('Cover return text for pic_url={}'.format(pic_url))
|
||||
return None
|
||||
pic_file.content_type = content_type
|
||||
default_storage.save(pic_file_name, pic_file)
|
||||
return default_storage.url(pic_file_name)
|
||||
except Exception, e:
|
||||
# if there is a problem, return None for cover URL
|
||||
logger.warning('Failed to store cover for username={}'.format(user))
|
||||
return None
|
||||
|
||||
|
||||
def selectively_associate_by_email(backend, details, user=None, *args, **kwargs):
|
||||
"""
|
||||
|
@ -30,25 +51,25 @@ def selectively_associate_by_email(backend, details, user=None, *args, **kwargs)
|
|||
return None
|
||||
return associate_by_email(backend, details, user=None, *args, **kwargs)
|
||||
|
||||
def facebook_extra_values( user, extra_data):
|
||||
def facebook_extra_values(user, extra_data):
|
||||
try:
|
||||
facebook_id = extra_data.get('id')
|
||||
user.profile.facebook_id = facebook_id
|
||||
profile_image_url = extra_data['picture']['data']['url']
|
||||
user.profile.pic_url = pic_storage_url(user, 'facebook', profile_image_url)
|
||||
if user.profile.avatar_source is None or user.profile.avatar_source is PRIVATETAR:
|
||||
user.profile.avatar_source = FACEBOOK
|
||||
user.profile.save()
|
||||
return True
|
||||
except Exception,e:
|
||||
logger.error(e)
|
||||
return False
|
||||
except Exception, e:
|
||||
logger.exception(e)
|
||||
return
|
||||
|
||||
def twitter_extra_values( user, extra_data):
|
||||
def twitter_extra_values(user, extra_data):
|
||||
try:
|
||||
twitter_id = extra_data.get('screen_name')
|
||||
profile_image_url = extra_data.get('profile_image_url_https')
|
||||
user.profile.twitter_id = twitter_id
|
||||
if user.profile.avatar_source is None or user.profile.avatar_source in (TWITTER, PRIVATETAR):
|
||||
user.profile.pic_url = profile_image_url
|
||||
user.profile.pic_url = pic_storage_url(user, 'twitter', profile_image_url)
|
||||
if user.profile.avatar_source is None or user.profile.avatar_source is PRIVATETAR:
|
||||
user.profile.avatar_source = TWITTER
|
||||
user.profile.save()
|
||||
|
@ -57,11 +78,11 @@ def twitter_extra_values( user, extra_data):
|
|||
logger.error(e)
|
||||
return False
|
||||
|
||||
def deliver_extra_data(backend, user, social, *args, **kwargs):
|
||||
def deliver_extra_data(backend, user, social, response, *args, **kwargs):
|
||||
if backend.name is 'twitter':
|
||||
twitter_extra_values( user, social.extra_data)
|
||||
if backend.name is 'facebook':
|
||||
facebook_extra_values( user, social.extra_data)
|
||||
facebook_extra_values( user, response)
|
||||
|
||||
# following is needed because of length limitations in a unique constrain for MySQL
|
||||
def chop_username(username, *args, **kwargs):
|
||||
|
@ -98,7 +119,7 @@ class SocialAuthExceptionMiddlewareWithoutMessages(SocialAuthExceptionMiddleware
|
|||
backend_name = getattr(backend, 'name', 'unknown-backend')
|
||||
|
||||
message = self.get_message(request, exception)
|
||||
social_logger.error(message)
|
||||
logger.warning(message)
|
||||
|
||||
url = self.get_redirect_uri(request, exception)
|
||||
url += ('?' in url and '&' or '?') + \
|
||||
|
|
|
@ -65,7 +65,7 @@ urlpatterns = [
|
|||
'password_reset_form': forms.SocialAwarePasswordResetForm},
|
||||
name='libraryauth_password_reset'),
|
||||
|
||||
url(r'^socialauth/', include('social.apps.django_app.urls', namespace='social')),
|
||||
url(r'^socialauth/', include('social_django.urls', namespace='social')),
|
||||
url('accounts/', include('email_change.urls')),
|
||||
url(r'^accounts/', include('registration.backends.model_activation.urls')),
|
||||
url(r'^accounts/', include('django.contrib.auth.urls')),
|
||||
|
|
|
@ -41,11 +41,11 @@ django-tastypie==0.13.3
|
|||
git+git://github.com/resulto/django-transmeta.git@ad4d7278ba330dcf8c8446f8ae9b2c769ae8684e
|
||||
fef-questionnaire==4.0.1
|
||||
#gitenberg.metadata==0.1.6
|
||||
git+https://github.com/gitenberg-dev/gitberg-build
|
||||
git+git://github.com/gitenberg-dev/gitberg-build.git@61a5fb0011e1a547b1eac14dd845ce37dbb5f85a
|
||||
#git+ssh://git@github.com/gitenberg-dev/metadata.git@0.1.11
|
||||
github3.py==0.9.5
|
||||
html5lib==1.0.1
|
||||
httplib2==0.7.5
|
||||
httplib2==0.11.3
|
||||
isodate==0.5.1
|
||||
kombu==3.0.35
|
||||
lxml==4.2.1
|
||||
|
@ -54,7 +54,7 @@ mechanize==0.2.5
|
|||
mimeparse==0.1.3
|
||||
nose==1.1.2
|
||||
numpy==1.11.2
|
||||
oauth2==1.5.211
|
||||
oauth2==1.9.0.post1
|
||||
oauthlib==1.1.2
|
||||
pandas==0.19.1
|
||||
paramiko==1.14.1
|
||||
|
@ -66,7 +66,8 @@ pyparsing==2.0.3
|
|||
python-dateutil==2.5.3
|
||||
python-mimeparse==0.1.4
|
||||
python-openid==2.2.5
|
||||
python-social-auth==0.2.21
|
||||
social-auth-core==1.7.0
|
||||
social-auth-app-django==2.1.0
|
||||
pytz==2016.6.1
|
||||
rdflib==4.2.0
|
||||
rdflib-jsonld==0.3
|
||||
|
|
|
@ -157,7 +157,7 @@ INSTALLED_APPS = (
|
|||
'regluit.payment',
|
||||
'regluit.utils',
|
||||
'registration',
|
||||
'social.apps.django_app.default',
|
||||
'social_django',
|
||||
'tastypie',
|
||||
'djcelery',
|
||||
'el_pagination',
|
||||
|
@ -243,72 +243,74 @@ SESSION_COOKIE_AGE = 3628800 # 6 weeks
|
|||
|
||||
# django-socialauth
|
||||
AUTHENTICATION_BACKENDS = (
|
||||
'social.backends.google.GoogleOAuth2',
|
||||
'social.backends.twitter.TwitterOAuth',
|
||||
'social.backends.yahoo.YahooOpenId',
|
||||
'social.backends.facebook.FacebookOAuth2',
|
||||
'social.backends.open_id.OpenIdAuth',
|
||||
'social_core.backends.google.GoogleOAuth2',
|
||||
'social_core.backends.twitter.TwitterOAuth',
|
||||
'social_core.backends.yahoo.YahooOpenId',
|
||||
'social_core.backends.facebook.FacebookOAuth2',
|
||||
'social_core.backends.open_id.OpenIdAuth',
|
||||
'django.contrib.auth.backends.ModelBackend',
|
||||
)
|
||||
|
||||
SOCIAL_AUTH_ENABLED_BACKENDS = ['google', 'facebook', 'twitter']
|
||||
#SOCIAL_AUTH_ASSOCIATE_BY_MAIL = True
|
||||
SOCIAL_AUTH_NEW_USER_REDIRECT_URL = '/'
|
||||
FACEBOOK_SOCIAL_AUTH_BACKEND_ERROR_URL = '/'
|
||||
SOCIAL_AUTH_SLUGIFY_USERNAMES = True
|
||||
SOCIAL_AUTH_NONCE_SERVER_URL_LENGTH = 200
|
||||
SOCIAL_AUTH_ASSOCIATION_SERVER_URL_LENGTH = 135
|
||||
SOCIAL_AUTH_ASSOCIATION_HANDLE_LENGTH = 125
|
||||
SOCIAL_AUTH_FACEBOOK_PROFILE_EXTRA_PARAMS = {'fields': 'picture'}
|
||||
SOCIAL_AUTH_FACEBOOK_LOGIN_ERROR_URL = '/'
|
||||
SOCIAL_AUTH_TWITTER_LOGIN_ERROR_URL = '/'
|
||||
|
||||
SOCIAL_AUTH_PIPELINE = (
|
||||
# Get the information we can about the user and return it in a simple
|
||||
# format to create the user instance later. On some cases the details are
|
||||
# already part of the auth response from the provider, but sometimes this
|
||||
# could hit a provider API.
|
||||
'social.pipeline.social_auth.social_details',
|
||||
'social_core.pipeline.social_auth.social_details',
|
||||
|
||||
# Get the social uid from whichever service we're authing thru. The uid is
|
||||
# the unique identifier of the given user in the provider.
|
||||
'social.pipeline.social_auth.social_uid',
|
||||
'social_core.pipeline.social_auth.social_uid',
|
||||
|
||||
# Verifies that the current auth process is valid within the current
|
||||
# project, this is were emails and domains whitelists are applied (if
|
||||
# defined).
|
||||
'social.pipeline.social_auth.auth_allowed',
|
||||
'social_core.pipeline.social_auth.auth_allowed',
|
||||
|
||||
# Checks if the current social-account is already associated in the site.
|
||||
'regluit.libraryauth.auth.selective_social_user',
|
||||
|
||||
# Make up a username for this person, appends a random string at the end if
|
||||
# there's any collision.
|
||||
'social.pipeline.user.get_username',
|
||||
'social_core.pipeline.user.get_username',
|
||||
|
||||
# make username < 222 in length
|
||||
'regluit.libraryauth.auth.chop_username',
|
||||
|
||||
# Send a validation email to the user to verify its email address.
|
||||
# Disabled by default.
|
||||
# 'social.pipeline.mail.mail_validation',
|
||||
# 'social_core.pipeline.mail.mail_validation',
|
||||
|
||||
# Associates the current social details with another user account with
|
||||
# a similar email address. don't use twitter or facebook to log in
|
||||
'regluit.libraryauth.auth.selectively_associate_by_email',
|
||||
|
||||
# Create a user account if we haven't found one yet.
|
||||
'social.pipeline.user.create_user',
|
||||
'social_core.pipeline.user.create_user',
|
||||
|
||||
# Create the record that associated the social account with this user.
|
||||
'social.pipeline.social_auth.associate_user',
|
||||
'social_core.pipeline.social_auth.associate_user',
|
||||
|
||||
# Populate the extra_data field in the social record with the values
|
||||
# specified by settings (and the default ones like access_token, etc).
|
||||
'social.pipeline.social_auth.load_extra_data',
|
||||
'social_core.pipeline.social_auth.load_extra_data',
|
||||
|
||||
# add extra data to user profile
|
||||
'regluit.libraryauth.auth.deliver_extra_data',
|
||||
|
||||
# Update the user record with any changed info from the auth service.
|
||||
'social.pipeline.user.user_details'
|
||||
'social_core.pipeline.user.user_details'
|
||||
)
|
||||
|
||||
SOCIAL_AUTH_TWITTER_EXTRA_DATA = [('profile_image_url_https', 'profile_image_url_https'),('screen_name','screen_name')]
|
||||
|
|
|
@ -22,6 +22,7 @@ EMAIL_HOST_PASSWORD = os.environ.get("EMAIL_HOST_PASSWORD", '012345678901234567
|
|||
# twitter auth
|
||||
# you'll need to create a new Twitter application to fill in these blanks
|
||||
# https://dev.twitter.com/apps/new
|
||||
# the field for redirect url must be filled in with https://unglue.it/socialauth/complete/twitter/?
|
||||
SOCIAL_AUTH_TWITTER_KEY = os.environ.get("SOCIAL_AUTH_TWITTER_KEY", '0123456789012345678901234')
|
||||
SOCIAL_AUTH_TWITTER_SECRET = os.environ.get("SOCIAL_AUTH_TWITTER_SECRET", '01234567890123456789012345678901234567890123456789')
|
||||
|
||||
|
|
Loading…
Reference in New Issue