text
stringlengths 226
34.5k
|
---|
Tkinter: Changing a variable within a function
Question: I know this kind of question gets asked all the time but either i've been
unable to come across the answer i need, or i've been unable to understand it
when i did.
I want to be able to do something like:
spam = StringVar()
spam.set(aValue)
class MyScale(Scale):
def __init__(self,var,*args,**kwargs):
Scale.__init__(self,*args,**kwargs)
self.bind("<ButtonRelease-1>",self.getValue)
self.set(var.get())
def getValue(self,event):
## spam gets changed to the new value set
## by the user manipulating the scale
var.set(self.get)
eggs = MyScale(spam,*args,**kwargs)
eggs.pack()
Of course, i get back "NameError: global name 'var' is not defined."
How do i get around the inability to pass arguments to getValue? I've been
warned against using global variables but is that my only option? Is it
setting up a separate scale class for each variable i want to change? I get
the feeling i'm missing something thats right under my nose...
edit: is this what you mean?
Exception in Tkinter callback
Traceback (most recent call last):
File "C:\Python26\lib\lib-tk\Tkinter.py", line 1410, in __call__
return self.func(*args)
File "C:\...\interface.py", line 70, in getValue
var.set(self.get)
NameError: global name 'var' is not defined
Sorry, I've only been programming a month and some of the jargon still escapes
me.
Answer: Please give this a shot.
Lots of example code out there generously uses globals, like your "var"
variable.
I have used your var argument to act as a pointer back to the original spam
object; assigned to self.var_pointer within the MyScale class.
The code below will change the value of 'spam' (and 'eggs') on the scale's
ButtonRelease.
You can check out the value by typing eggs.get() or spam.get() to see the
changed value.
from Tkinter import *
root = Tk()
aValue = "5"
spam = StringVar()
spam.set(aValue)
class MyScale(Scale):
def __init__(self,var,*args,**kwargs):
self.var_pointer = var
Scale.__init__(self,*args,**kwargs)
self.bind("<ButtonRelease-1>",self.getValue)
self.set(var.get())
def getValue(self,event):
## spam gets changed to the new value set
## by the user manipulating the scale
self.var_pointer.set(self.get())
eggs = MyScale(spam)
eggs.pack(anchor=CENTER)
|
Different styles for Windows forms in Ironpython
Question: I want to change the look of my Ironpython windows forms, Is it possible to
change the style of the form and for example make it more like a Mac?
thank you
Answer: As an interface designer, it's important to use an many standard windows
controls as possible. The more your app looks like the microsoft ones users
are used to, they will become more comfortable with it quicker.
A good example is how out of place iTunes and Safari look in windows. My tip
is to just keep it with the windows look.
|
Write to UTF-8 file in Python
Question: I'm really confused with the `codecs.open function`. When I do:
file = codecs.open("temp", "w", "utf-8")
file.write(codecs.BOM_UTF8)
file.close()
It gives me the error
> UnicodeDecodeError: 'ascii' codec can't decode byte 0xef in position 0:
> ordinal not in range(128)
If I do:
file = open("temp", "w")
file.write(codecs.BOM_UTF8)
file.close()
It works fine.
**Question** is why does the first method fail? And how do I insert the bom?
If the second method is the correct way of doing it, what the point of using
`codecs.open(filename, "w", "utf-8")`?
Answer: Disclaimer: I'm not a Python programmer.
I believe the problem is that
[`codecs.BOM_UTF8`](https://docs.python.org/library/codecs.html#codecs.BOM_UTF8)
is a byte string, not a Unicode string. I suspect the file handler is trying
to guess what you really mean based on "I'm meant to be writing Unicode as
UTF-8-encoded text, but you've given me a byte string!"
Try writing the Unicode string for the byte order mark (i.e. Unicode U+FEFF)
directly, so that the file just encodes that as UTF-8:
import codecs
file = codecs.open("lol", "w", "utf-8")
file.write(u'\ufeff')
file.close()
(That seems to give the right answer - a file with bytes EF BB BF.)
EDIT: S. Lott's [suggestion](http://stackoverflow.com/a/934203/12892) of using
"utf-8-sig" as the encoding is a better one than explicitly writing the BOM
yourself, but I'll leave this answer here as it explains what was going wrong
before.
|
Creating a logging handler to connect to Oracle?
Question: So right now i need to create and implement an extension of the Python logging
module that will be used to log to our database. Basically we have several
python applications(that all run in the background) that currently log to a
random mishmash of text files. Which makes it almost impossible to find out if
a certain application failed or not.
The problem given to me is to move said logging to text files to an oracle DB.
The tables have already been defined, and where things need to be logged to
but right now, im looking at adding another logging handler that will log to
the DB.
I am using python 2.5.4 and cx_Oracle and the applications in general can be
ether run as a service/daemon or a straight application.
I'm just mainly curious about what would be the best possible way to go about
this. Few questions:
1. If any errors occur with cx_Oracle, where should these errors be logged to? If its down would it be best to just go and have the logger retreat to the default text file?
2. Awhile back we started enforcing that people use sys.stderr/stdout.write instead of print, so worst case scenario we wouldn't run into any issues with print becoming deprecated. Is there a way to seamlessly make all of the thousands of sys.std calls be piped directly into the logger, and have the logger pickup the slack?
3. After every logged message, should the script automatically do a commit? (there's going to be several dozen a second.)
4. What is the best way to implement a new handler for the logging system? Inheriting from the basic Handler class seems to be easiest.
Any ideas / suggestions would be great.
Answer: 1. If errors occur with cx_Oracle, it's probably best to log these to a text file.
2. You could try redirecting sys.stdout and sys.stderr to file-like objects which log whatever's written to them to a logger.
3. I would guess you do want to commit after each event, unless you have strong reasons for not doing this. Alternatively, you can buffer several events and write them all in a single transaction every so often.
4. Below is an example which uses mx.ODBC, you can probably adapt this to cx_Oracle without too much trouble. It's meant to be Python DB-API 2.0 compliant, I think.
The standalone Python logging distribution (before logging was added to
Python) is at <http://www.red-dove.com/python_logging.html> and although the
logging package in Python is much more up to date, the standalone distribution
contains a test directory which has a lot of useful examples of derived
handler classes.
#!/usr/bin/env python
#
# Copyright 2001-2009 by Vinay Sajip. All Rights Reserved.
#
# Permission to use, copy, modify, and distribute this software and its
# documentation for any purpose and without fee is hereby granted,
# provided that the above copyright notice appear in all copies and that
# both that copyright notice and this permission notice appear in
# supporting documentation, and that the name of Vinay Sajip
# not be used in advertising or publicity pertaining to distribution
# of the software without specific, written prior permission.
# VINAY SAJIP DISCLAIMS ALL WARRANTIES WITH REGARD TO THIS SOFTWARE, INCLUDING
# ALL IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL
# VINAY SAJIP BE LIABLE FOR ANY SPECIAL, INDIRECT OR CONSEQUENTIAL DAMAGES OR
# ANY DAMAGES WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER
# IN AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT
# OF OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
#
# This file is part of the standalone Python logging distribution. See
# http://www.red-dove.com/python_logging.html
#
"""
A test harness for the logging module. An example handler - DBHandler -
which writes to an Python DB API 2.0 data source. You'll need to set this
source up before you run the test.
Copyright (C) 2001-2009 Vinay Sajip. All Rights Reserved.
"""
import sys, string, time, logging
class DBHandler(logging.Handler):
def __init__(self, dsn, uid='', pwd=''):
logging.Handler.__init__(self)
import mx.ODBC.Windows
self.dsn = dsn
self.uid = uid
self.pwd = pwd
self.conn = mx.ODBC.Windows.connect(self.dsn, self.uid, self.pwd)
self.SQL = """INSERT INTO Events (
Created,
RelativeCreated,
Name,
LogLevel,
LevelText,
Message,
Filename,
Pathname,
Lineno,
Milliseconds,
Exception,
Thread
)
VALUES (
%(dbtime)s,
%(relativeCreated)d,
'%(name)s',
%(levelno)d,
'%(levelname)s',
'%(message)s',
'%(filename)s',
'%(pathname)s',
%(lineno)d,
%(msecs)d,
'%(exc_text)s',
'%(thread)s'
);
"""
self.cursor = self.conn.cursor()
def formatDBTime(self, record):
record.dbtime = time.strftime("#%m/%d/%Y#", time.localtime(record.created))
def emit(self, record):
try:
#use default formatting
self.format(record)
#now set the database time up
self.formatDBTime(record)
if record.exc_info:
record.exc_text = logging._defaultFormatter.formatException(record.exc_info)
else:
record.exc_text = ""
sql = self.SQL % record.__dict__
self.cursor.execute(sql)
self.conn.commit()
except:
import traceback
ei = sys.exc_info()
traceback.print_exception(ei[0], ei[1], ei[2], None, sys.stderr)
del ei
def close(self):
self.cursor.close()
self.conn.close()
logging.Handler.close(self)
dh = DBHandler('Logging')
logger = logging.getLogger("")
logger.setLevel(logging.DEBUG)
logger.addHandler(dh)
logger.info("Jackdaws love my big %s of %s", "sphinx", "quartz")
logger.debug("Pack my %s with five dozen %s", "box", "liquor jugs")
try:
import math
math.exp(1000)
except:
logger.exception("Problem with %s", "math.exp")
|
How to Change Mouse Cursor in PythonCard
Question: How do I change the mouse cursor to indicate a waiting state using Python and
PythonCard?
I didn't see anything in the documentation.
Answer: PythonCard builds on top of wx, so if you import wx you should be able to
build a suitable cursor (e.g. with `wx.CursorFromImage`), set it (e.g. with
`wx.BeginBusyCursor`) when your wait begins, and end it (with
`wx.EndBusyCursor`) when your wait ends.
|
How to check for NaN in python?
Question: `float('nan')` results in a thingy simply called nan. But how do I check for
it? Should be very easy, but i cannot find it.
Answer: [math.isnan()](http://docs.python.org/library/math.html#math.isnan)
> Checks if the float x is a NaN (not a number). NaNs are part of the IEEE 754
> standards. Operation like but not limited to inf * 0, inf / inf or any
> operation involving a NaN, e.g. nan * 1, return a NaN.
>
> _New in version 2.6._
>>> import math
>>> x=float('nan')
>>> math.isnan(x)
True
>>>
|
Using Python's list index() method on a list of tuples or objects?
Question: Python's list type has an index() method that takes one parameter and returns
the index of the first item in the list matching the parameter. For instance:
>>> some_list = ["apple", "pear", "banana", "grape"]
>>> some_list.index("pear")
1
>>> some_list.index("grape")
3
Is there a graceful (idiomatic) way to extend this to lists of complex
objects, like tuples? Ideally, I'd like to be able to do something like this:
>>> tuple_list = [("pineapple", 5), ("cherry", 7), ("kumquat", 3), ("plum", 11)]
>>> some_list.getIndexOfTuple(1, 7)
1
>>> some_list.getIndexOfTuple(0, "kumquat")
2
getIndexOfTuple() is just a hypothetical method that accepts a sub-index and a
value, and then returns the index of the list item with the given value at
that sub-index. I hope
Is there some way to achieve that general result, using list comprehensions or
lambas or something "in-line" like that? I think I could write my own class
and method, but I don't want to reinvent the wheel if Python already has a way
to do it.
Answer: How about this?
>>> tuple_list = [("pineapple", 5), ("cherry", 7), ("kumquat", 3), ("plum", 11)]
>>> [x for x, y in enumerate(tuple_list) if y[1] == 7]
[1]
>>> [x for x, y in enumerate(tuple_list) if y[0] == 'kumquat']
[2]
As pointed out in the comments, this would get all matches. To just get the
first one, you can do:
>>> [y[0] for y in tuple_list].index('kumquat')
2
There is a good discussion in the comments as to the speed difference between
all the solutions posted. I may be a little biased but I would personally
stick to a one-liner as the speed we're talking about is pretty insignificant
versus creating functions and importing modules for this problem, but if you
are planning on doing this to a very large amount of elements you might want
to look at the other answers provided, as they are faster than what I
provided.
|
Why won't python allow me to delete files?
Question: I've created a python script that gets a list of files from a text file and
deletes them if they're empty. It correctly detects empty files but it doesn't
want to delete them. It gives me:
(32, 'The process cannot access the file because it is being used by another process')
I've used two different tools to check whether the files are locked or not and
I'm certain that they are not. I used sysinternals process explorer and
LockHunter. Furthermore, I'm able to just manually delete the files myself. I
obviously don't want to do that for all of them as there are hundreds in
various locations.
The script:
import os.path
import sys
def DeleteFilesFromListIfBlank(PathToListOfFiles):
ListOfFiles = open(PathToListOfFiles)
FilesToCheck = [];
for line in ListOfFiles.readlines():
if(len(line) > 1):
line = line.rstrip();
FilesToCheck.append(line)
print "Found %s files to check. Starting check." % len(FilesToCheck)
FilesToRemove = [];
for line in FilesToCheck:
#print "Opening %s" % line
try:
ActiveFile = open(line);
Length = len(ActiveFile.read())
if(Length < 691 and ActiveFile.read() == ""):
print "Deleting %s" % line
os.unlink(line);
else:
print "Keeping %s" % line
except IOError,message:
print "Could not open file: $s" % message
except Exception as inst:
print inst.args
DeleteFilesFromListIfBlank("C:\\ListOfResx.txt")
I've tried using both os.unlink and os.remove. I'm running Python 2.6 on
Vista64
Thanks
Answer: You need to call `.close()` on the file object before you try and delete it.
Edit: And really you shouldn't be opening the file at all. `os.stat()` will
tell you the size of a file (and 9 other values) without ever opening the
file.
This (I think) does the same thing but is a little cleaner (IMHO):
import os
_MAX_SIZE = 691
def delete_if_blank(listFile):
# Make a list of files to check.
with open(listFile) as listFile:
filesToCheck = filter(None, (line.rstrip() for line in listFile.readlines()))
# listFile is automatically closed now because we're out of the 'with' statement.
print "Found %u files to check. Starting check." % len(filesToCheck)
# Remove each file.
for filename in filesToCheck:
if os.stat(filename).st_size < _MAX_SIZE:
print "Deleting %s" % filename
os.remove(filename)
else:
print "Keeping %s" % filename
|
Prevent a timer from updating a text box if the key cursor is in the box
Question: Is it possible to check if a
[`TextCtrl`](http://www.wxpython.org/docs/api/wx.TextCtrl-class.html) is under
keyboard focus (blinking cursor in text box) without defining a handler for
[`EVT_SET_FOCUS`](http://www.wxpython.org/docs/api/wx.FocusEvent-class.html)?
I just want to do a quick boolean check to prevent a `wx.Timer` from
overwriting the text box if the user is writing something in the box.
Answer: You can bypass a timer update by finding which window has the focus (using
`FindFocus`) and comparing this to your TextCtrl window. Then, if your
TextCtrl has the focus you can leave it alone. Here's an example:
import wx
class TestFrame(wx.Frame):
def __init__(self):
self.count = 0
wx.Frame.__init__(self, None, -1, "test frame", size=(200, 100))
self.panel = wx.Panel(self, -1)
button = wx.Button(self.panel, -1, "b", pos=(10, 40))
self.text = wx.TextCtrl(self.panel, -1, `self.count`, size=(50, 25))
self.timer = wx.Timer(self)
self.Bind(wx.EVT_TIMER, self.incr_text)
self.timer.Start(1000)
def incr_text(self, event):
self.count += 1
f = self.FindFocus()
if not f==self.text:
self.text.SetValue(`self.count`)
if __name__=="__main__":
app = wx.PySimpleApp()
TestFrame().Show()
app.MainLoop()
|
Python - simple reading lines from a pipe
Question: I'm trying to read lines from a pipe and process them, but I'm doing something
silly and I can't figure out what. The producer is going to keep producing
lines indefinitely, like this:
producer.py
import time
while True:
print 'Data'
time.sleep(1)
The consumer just needs to check for lines periodically:
consumer.py
import sys, time
while True:
line = sys.stdin.readline()
if line:
print 'Got data:', line
else:
time.sleep(1)
When I run this in the Windows shell as `python producer.py | python consumer.py`, it just sleeps forever (never seems to get data?) It seems that maybe the problem is that the producer never terminates, since if I send a finite amount of data then it works fine.
How can I get the data to be received and show up for the consumer? In the
real application, the producer is a C++ program I have no control over.
Answer: Some old versions of Windows simulated pipes through files (so they were prone
to such problems), but that hasn't been a problem in 10+ years. Try adding a
sys.stdout.flush()
to the producer after the `print`, and also try to make the producer's stdout
unbuffered (by using `python -u`).
Of course this doesn't help if you have no control over the producer -- if it
buffers too much of its output you're still going to wait a long time.
Unfortunately - while there are many approaches to solve that problem on Unix-
like operating systems, such as pyexpect,
[pexpect](http://sourceforge.net/projects/pexpect/),
[exscript](http://code.google.com/p/exscript/), and
[paramiko](http://www.lag.net/paramiko/), I doubt any of them works on
Windows; if that's indeed the case, I'd try [Cygwin](http://www.cygwin.com/),
which puts enough of a Linux-like veneer on Windows as to often enable the use
of Linux-like approaches on a Windows box.
|
cx_Oracle And User Defined Types
Question: Does anyone know an easier way to work with user defined types in Oracle using
cx_Oracle?
For example, if I have these two types:
CREATE type my_type as object(
component varchar2(30)
,key varchar2(100)
,value varchar2(4000))
/
CREATE type my_type_tab as table of my_type
/
And then a procedure in package `my_package` as follows:
PROCEDURE my_procedure (param in my_type_tab);
To execute the procedure in PL/SQL I can do something like this:
declare
l_parms my_type_tab;
l_cnt pls_integer;
begin
l_parms := my_type_tab();
l_parms.extend;
l_cnt := l_parms.count;
l_parms(l_cnt) := my_type('foo','bar','hello');
l_parms.extend;
l_cnt := l_parms.count;
l_parms(l_cnt) := my_type('faz','baz','world');
my_package.my_procedure(l_parms);
end;
However, I was wondering how I can do it in Python, similar to this code:
import cx_Oracle
orcl = cx_Oracle.connect('foo:[email protected]:5555/blah' + instance)
curs = orcl.cursor()
params = ???
curs.execute('begin my_package.my_procedure(:params)', params=params)
If the parameter was a string I can do this as above, but since it's an user-
defined type, I have no idea how to call it without resorting to pure PL/SQL
code.
Edit: Sorry, I should have said that I was looking for ways to do more in
Python code instead of PL/SQL.
Answer: While cx_Oracle can select user defined types, it does not to my knowledge
support passing in user defined types as bind variables. So for example the
following will work:
cursor.execute("select my_type('foo', 'bar', 'hello') from dual")
val, = cursor.fetchone()
print val.COMPONENT, val.KEY, val.VALUE
However what you can't do is construct a Python object, pass it in as an input
argument and then have cx_Oracle "translate" the Python object into your
Oracle type. So I would say you're going to have to construct your input
argument within a PL/SQL block.
You can pass in Python lists, so the following should work:
components=["foo", "faz"]
values=["bar", "baz"]
keys=["hello", "world"]
cursor.execute("""
declare
type udt_StringList is table of varchar2(4000) index by binary_integer;
l_components udt_StringList := :p_components;
l_keys udt_StringList := :p_keys;
l_values udt_StringList := :p_values;
l_parms my_type_tab;
begin
l_parms.extend(l_components.count);
for i in 1..l_components.count loop
l_parms(i) := my_type(l_components(i), l_keys(i), l_values(i));
end loop;
my_package.my_procedure(l_parms);
end;""", p_components=components, p_values=values, p_keys=keys)
|
Redirecting FORTRAN (called via F2PY) output in Python
Question: I'm trying to figure out how to redirect output from some FORTRAN code for
which I've generated a Python interface by using F2PY. I've tried:
from fortran_code import fortran_function
stdout_holder = sys.stdout
stderr_holder = sys.stderr
sys.stdout = file("/dev/null","w")
fortran_function()
sys.stdout.close()
sys.stderr.close()
sys.stdout = stdout_holder
sys.stderr = stderr_holder
This is the de facto method of redirecting output in Python, but it doesn't
seem to work in this case (i.e., the output is displayed anyway).
I did find [a mailing list post from 2002](http://cens.ioc.ee/pipermail/f2py-
users/2002-November/000352.html) saying that "It is possible to read messages
from pts devices, e.g. ttysnoop does this". Information on ttysnoop seems to
be pretty difficult to find online (I don't think it's been updated in quite a
few years; for example, [the first result on Google for
"ttysnoop"](http://www.linuxhelp.net/guides/ttysnoop/) has only dead links to
tarballs, RPMs, and .deb's), and [this request for a port to OS
X](http://macosx.forked.net/bbarchive/forum5/000008.html) received the
response "No luck, it requires some linux specific utmp functions which I
can't create."
I'm open to any suggestions on how to redirect the output (it doesn't have to
use ttysnoop).
Thanks!
Answer: The stdin and stdout fds are being inherited by the C shared library.
from fortran_code import fortran_function
import os
print "will run fortran function!"
# open 2 fds
null_fds = [os.open(os.devnull, os.O_RDWR) for x in xrange(2)]
# save the current file descriptors to a tuple
save = os.dup(1), os.dup(2)
# put /dev/null fds on 1 and 2
os.dup2(null_fds[0], 1)
os.dup2(null_fds[1], 2)
# *** run the function ***
fortran_function()
# restore file descriptors so I can print the results
os.dup2(save[0], 1)
os.dup2(save[1], 2)
# close the temporary fds
os.close(null_fds[0])
os.close(null_fds[1])
print "done!"
|
Obfuscate strings in Python
Question: I have a password string that must be passed to a method. Everything works
fine but I don't feel comfortable storing the password in clear text. Is there
a way to obfuscate the string or to truly encrypt it? I'm aware that
obfuscation can be reverse engineered, but I think I should at least try to
cover up the password a bit. At the very least it wont be visible to a
indexing program, or a stray eye giving a quick look at my code.
I am aware of pyobfuscate but I don't want the whole program obfuscated, just
one string and possibly the whole line itself where the variable is defined.
Target platform is GNU Linux Generic (If that makes a difference)
Answer: If you just want to prevent casually glancing at a password, you may want to
consider encoding/decoding the password to/from
[base64](http://docs.python.org/library/base64.html). It's not secure in the
least, but the password won't be casually human/robot readable.
import base64
# Encode password
encoded_pw = base64.b64encode(raw_pw)
# Decode password
decoded_pw = base64.b64decode(encoded_pw)
|
Python 3 doesn't read unicode file on a new server
Question: My webpages are served by a script that dynamically imports a bunch of files
with
try:
with open (filename, 'r') as f:
exec(f.read())
except IOError: pass
(actually, can you suggest a better method of importing a file? I'm sure there
is one.)
Sometimes the files have strings in different languages, like
# contents of language.ru
title = "Название"
Those were all saved as UTF-8 files. Python has no problem running the script
in command line or serving a page from my MacBook:
OK: [server command line] python3.0 page.py /index.ru
OK: http://whitebox.local/index.ru
but it throws an error when trying to serve a page from a server we just moved
to:
157 try:
158 with open (filename, 'r') as f:
159 exec(f.read())
160 except IOError: pass
161
/usr/local/lib/python3.0/io.py in read(self=, n=-1)
...
UnicodeDecodeError: 'ascii' codec can't decode byte 0xe1 in position 627: ordinal not in range(128)
All the files were copied from my laptop where they were perfectly served by
Apache. What is the reason?
**Update:** I found out the default encoding for `open()` is platform-
dependent so it was `utf8` on my laptop and `ascii` on server. I wonder if
there is a per-program function to set it in Python 3
(`sys.setdefaultencoding` is used in `site` module and then deleted from the
namespace).
Answer: Use `open(filename, 'r', encoding='utf8')`. See [Python 3 docs for
open](http://docs.python.org/3.0/library/functions.html#open).
|
Django : Timestamp string custom field
Question: I'm trying to create a custom timestamp field.
class TimestampKey(models.CharField):
__metaclass__ = models.SubfieldBase
def __init__(self, *args, **kwargs):
import time
kwargs['unique'] = True
kwargs['max_length'] = 20
kwargs['auto_created'] = True
kwargs['editable']=False
super(TimestampKey, self).__init__(*args, **kwargs)
def to_python(self, value) :
return value
def get_db_prep_value(self, value) :
try:
import time
t = time.localtime()
value = reduce(lambda a,b:str(a)+str(b),t)
except ValueError:
value = {}
return value
class Table1(models.Model):
f = TimestampKey(primary_key=True)
n = ....
It stores the value with appropriate timestamp in the db. But it doesnt
populate the field 'f' in the object.
Eg:
t1 = Table1(n="some value")
t1.f -> blank
t1.save()
t1.f -> blank.
This is the problem. Am I missing something so that it doesnt populate the
filed? Please shed some light on this.
Thanks.
Answer: Is it wise to use a timestamp as your primary key? If your database uses ISO
8601 or really any time format in which second is the smallest time
interval... Well, anyway, my point is that you have no guarantee, especially
if this is going to be a web-facing application that two entries are going to
resolve within the minimum time interval. That is, if the smallest time
interval is a second, as in ISO 8601, if you get two requests to save in the
same second, you're going to get an error condition. Why not stick to
automatically incrementing integer keys and just make the timestamp its own
field?
|
an error in taking an input in python
Question: 111111111111111111111111111111111111111111111111111111111111
when i take this as input , it appends an L at the end like this
111111111111111111111111111111111111111111111111111111111111L
thus affecting my calculations on it .. how can i remove it?
import math
t=raw_input()
l1=[]
a=0
while (str(t)!="" and int(t)!= 0):
l=1
k=int(t)
while(k!= 1):
l=l+1
a=(0.5 + 2.5*(k %2))*k + k % 2
k=a
l1.append(l)
t=raw_input()
a=a+1
for i in range(0,int(a)):
print l1[i]
this is my code and it works for every test case except
111111111111111111111111111111111111111111111111111111111111
so i guess something is wrong when python considers such a huge number
Answer: It's being input as a Long Integer, which should behave just like any other
number in terms of doing calculations. It's only when you display it using
[`repr`](http://docs.python.org/library/functions.html#repr) (or something
that invokes `repr`, like printing a list) that it gets the 'L'.
What exactly is going wrong?
**Edit** : Thanks for the code. As far as I can see, giving it a long or short
number makes no difference, but it's not really clear what it's supposed to
do.
|
Counting repeated characters in a string in Python
Question: I want to count the number of times each character is repeated in a string. Is
there any particular way to do it apart from comparing each character of the
string from A-Z and incrementing a counter?
**Update** (in reference to [Anthony's
answer](http://stackoverflow.com/questions/991350/counting-repeated-
characters-in-a-string-in-python/991372#991372)): Whatever you have suggested
till now I have to write 26 times. Is there an easier way?
Answer:
import collections
d = collections.defaultdict(int)
for c in thestring:
d[c] += 1
A `collections.defaultdict` is like a `dict` (subclasses it, actually), but
when an entry is sought and not found, instead of reporting it doesn't have
it, it makes it and inserts it by calling the supplied 0-argument callable.
Most popular are `defaultdict(int)`, for counting (or, equivalently, to make a
multiset AKA bag data structure), and `defaultdict(list)`, which does away
forever with the need to use `.setdefault(akey, []).append(avalue)` and
similar awkward idioms.
So once you've done this `d` is a dict-like container mapping every character
to the number of times it appears, and you can emit it any way you like, of
course. For example, most-popular character first:
for c in sorted(d, key=d.get, reverse=True):
print '%s %6d' % (c, d[c])
|
Pylons FormEncode with an array of form elements
Question: I have a Pylons app and am using FormEncode and HtmlFill to handle my forms. I
have an array of text fields in my template (Mako)
<tr>
<td>Yardage</td>
<td>${h.text('yardage[]', maxlength=3, size=3)}</td>
<td>${h.text('yardage[]', maxlength=3, size=3)}</td>
<td>${h.text('yardage[]', maxlength=3, size=3)}</td>
<td>${h.text('yardage[]', maxlength=3, size=3)}</td>
<td>${h.text('yardage[]', maxlength=3, size=3)}</td>
<td>${h.text('yardage[]', maxlength=3, size=3)}</td>
<td>${h.text('yardage[]', maxlength=3, size=3)}</td>
<td>${h.text('yardage[]', maxlength=3, size=3)}</td>
<td>${h.text('yardage[]', maxlength=3, size=3)}</td>
</tr>
However, I can't seem to figure out how to validate these fields. Here is the
relevant entry from my Schema
`yardage = formencode.ForEach(formencode.validators.Int())`
I'm trying to validate that each of these fields is an Int. However, no
validation occurs for these fields.
**UPDATE** As requested here is the code for the action of this controller. I
know it was working as I can validate other form fields.
def submit(self):
schema = CourseForm()
try:
c.form_result = schema.to_python(dict(request.params))
except formencode.Invalid, error:
c.form_result = error.value
c.form_errors = error.error_dict or {}
c.heading = 'Add a course'
html = render('/derived/course/add.html')
return htmlfill.render(
html,
defaults = c.form_result,
errors = c.form_errors
)
else:
h.redirect_to(controler='course', action='view')
**UPDATE** It was suggested on IRC that I change the name of the elements from
`yardage[]` to `yardage` No result. They should all be ints but putting in f
into one of the elements doesn't cause it to be invalid. As I said before, I
am able to validate other form fields. Below is my entire schema.
import formencode
class CourseForm(formencode.Schema):
allow_extra_fields = True
filter_extra_fields = True
name = formencode.validators.NotEmpty(messages={'empty': 'Name must not be empty'})
par = formencode.ForEach(formencode.validators.Int())
yardage = formencode.ForEach(formencode.validators.Int())
Answer: Turns out what I wanted to do wasn't quite right.
**Template** :
<tr>
<td>Yardage</td>
% for hole in range(9):
<td>${h.text('hole-%s.yardage'%(hole), maxlength=3, size=3)}</td>
% endfor
</tr>
(Should have made it in a loop to begin with.) You'll notice that the name of
the first element will become `hole-1.yardage`. I will then use
`[FormEncode.variabledecode](http://www.formencode.org/en/latest/modules/variabledecode.html)`
to turn this into a dictionary. This is done in the
**Schema** :
import formencode
class HoleSchema(formencode.Schema):
allow_extra_fields = False
yardage = formencode.validators.Int(not_empty=True)
par = formencode.validators.Int(not_empty=True)
class CourseForm(formencode.Schema):
allow_extra_fields = True
filter_extra_fields = True
name = formencode.validators.NotEmpty(messages={'empty': 'Name must not be empty'})
hole = formencode.ForEach(HoleSchema())
The HoleSchema will validate that `hole-#.par` and `hole-#.yardage` are both
ints and are not empty. `formencode.ForEach` allows me to apply `HoleSchema`
to the dictionary that I get from passing `variable_decode=True` to the
`@validate` decorator.
Here is the `submit` action from my
**Controller** :
@validate(schema=CourseForm(), form='add', post_only=False, on_get=True,
auto_error_formatter=custom_formatter,
variable_decode=True)
def submit(self):
# Do whatever here.
return 'Submitted!'
Using the `@validate` decorator allows for a much cleaner way to validate and
fill in the forms. The `variable_decode=True` is very important or the
dictionary will not be properly created.
|
Segmentation fault in custom QAbstractItemModel
Question: I've written my own QAbstractItemModel to show a tree in TreeView. It shows
the top level items, but when you expand a directory, the app closes, the the
following message is written to the console: "Segmentation fault" What am I
doing wrong that is causing this. Here is a simplifed version of my code:
#!/usr/bin/env python
import sys
from PyQt4 import QtCore, QtGui
class TreeModel(QtCore.QAbstractItemModel):
NAME = 0
FILEID = QtCore.Qt.UserRole + 1
horizontalHeaderLabels = ["File Name",]
inventory = None
def set_tree(self, inventory, root_item):
self.emit(QtCore.SIGNAL("layoutAboutToBeChanged()"))
self.inventory = inventory
self.id2fileid = []
self.fileid2id = {}
self.dir_children_ids = {}
self.parent_ids = []
# Create internal ids for all items in the tree for use in
# ModelIndex's.
root_fileid = root_item.file_id
self.append_fileid(root_fileid, None)
remaining_dirs = [root_fileid,]
while remaining_dirs:
dir_fileid = remaining_dirs.pop(0)
dir_id = self.fileid2id[dir_fileid]
dir_children_ids = []
for child in inventory[dir_fileid].children:
id = self.append_fileid(child.file_id, dir_id)
dir_children_ids.append(id)
if child.children:
remaining_dirs.append(child.file_id)
if len(self.id2fileid) % 100 == 0:
QtCore.QCoreApplication.processEvents()
self.dir_children_ids[dir_id] = dir_children_ids
self.emit(QtCore.SIGNAL("layoutChanged()"))
def append_fileid(self, fileid, parent_id):
ix = len(self.id2fileid)
self.id2fileid.append(fileid)
self.parent_ids.append(parent_id)
self.fileid2id[fileid] = ix
return ix
def columnCount(self, parent):
if parent.isValid():
return 0
return len(self.horizontalHeaderLabels)
def rowCount(self, parent):
if self.inventory is None:
return 0
parent_id = parent.internalId()
if parent_id not in self.dir_children_ids:
return 0
return len(self.dir_children_ids[parent_id])
def _index(self, row, column, parent_id):
item_id = self.dir_children_ids[parent_id][row]
return self.createIndex(row, column, item_id)
def index(self, row, column, parent = QtCore.QModelIndex()):
if self.inventory is None:
return self.createIndex(row, column, 0)
parent_id = parent.internalId()
return self._index(row, column, parent_id)
def sibling(self, row, column, index):
sibling_id = child.internalId()
if sibling_id == 0:
return QtCore.QModelIndex()
parent_id = self.parent_ids[child_id]
return self._index(row, column, parent_id)
def parent(self, child):
child_id = child.internalId()
if child_id == 0:
return QtCore.QModelIndex()
item_id = self.parent_ids[child_id]
if item_id == 0 :
return self.createIndex(0, 0, item_id)
parent_id = self.parent_ids[item_id]
row = self.dir_children_ids[parent_id].index(item_id)
return self.createIndex(row, 0, item_id)
def hasChildren(self, parent):
if self.inventory is None:
return False
parent_id = parent.internalId()
return parent_id in self.dir_children_ids
def data(self, index, role):
if not index.isValid():
return QtCore.QVariant()
fileid = self.id2fileid[index.internalId()]
if role == self.FILEID:
return QtCore.QVariant(fileid)
item = self.inventory[fileid]
column = index.column()
if column == self.NAME:
if role == QtCore.Qt.DisplayRole:
return QtCore.QVariant(item.file_name)
return QtCore.QVariant()
def flags(self, index):
if not index.isValid():
return QtCore.Qt.ItemIsEnabled
return QtCore.Qt.ItemIsEnabled | QtCore.Qt.ItemIsSelectable
def headerData(self, section, orientation, role):
if orientation == QtCore.Qt.Horizontal and role == QtCore.Qt.DisplayRole:
return QtCore.QVariant(self.horizontalHeaderLabels[section])
return QtCore.QVariant()
inventory = {}
class InventoryItem():
def __init__(self, file_id, file_name, children=[]):
self.file_id = file_id
self.file_name = file_name
self.children = children
global inventory
inventory[file_id] = self
root_item = InventoryItem("root-id", "", [
InventoryItem("dir1-id", "dir1", [
InventoryItem("file1-id", "file1")
]),
InventoryItem("file1-id", "file1")
])
app = QtGui.QApplication(sys.argv)
model = TreeModel()
model.set_tree(inventory, root_item)
tree_view = QtGui.QTreeView()
tree_view.setModel(model)
tree_view.show()
app.exec_()
The full version can be found in this branch:
https://code.launchpad.net/~garyvdm/qbzr/trees, in the file lib/browse.py
Answer: Use the modeltest.py module! It exercises your model for different scenarios.
You can find it in the /contrib/ directory inside the PyQt source code
package.
|
Python - Print on stdout on a "terminal"
Question: Before starting, I ask you all to apologize for the question. Maybe it is
stupid, but I cannot find a solution. I am working on a remote machine, and
have no idea what type.
My python code, that seems to work, is the one below. The problem is that I am
trying to print some outputs on the screen but nothing happens. I have tried
both print and raw_input but nothing happens ... Do you know any other way to
do it ?
# Set up fields of reply message based on query
def prepareReply():
global authorReply, authorReplyLen, localConvId, originConvId, blbContentAndUntUnz, linkName
print "PLOP!"
raw_input("blabla")
#print "="*10
Thanks !
Answer:
import sys
print "Hi!"
sys.stdout.flush()
|
Command Line Arguments In Python
Question: I am originally a C programmer. I have seen numerous tricks and "hacks" to
read many different arguments.
What are some of the ways Python programmers can do this?
### Related
* [What’s the best way to grab/parse command line arguments passed to a Python script?](http://stackoverflow.com/questions/20063/whats-the-best-way-to-grab-parse-command-line-arguments-passed-to-a-python-scrip)
* [Implementing a “[command] [action] [parameter]” style command-line interfaces?](http://stackoverflow.com/questions/362426/implementing-a-command-action-parameter-style-command-line-interfaces)
* [How can I process command line arguments in Python?](http://stackoverflow.com/questions/567879/how-can-i-process-command-line-arguments-in-python)
* [How do I format positional argument help using Python’s optparse?](http://stackoverflow.com/questions/642648/how-do-i-format-positional-argument-help-using-pythons-optparse)
Answer:
import sys
print "\n".join(sys.argv)
`sys.argv` is a list that contains all the arguments passed to the script on
the command line.
Basically,
import sys
print sys.argv[1:]
|
Bash or Python to go backwards?
Question: I have a text file which a lot of random occurrences of the string @STRING_A,
and I would be interested in writing a short script which removes only some of
them. Particularly one that scans the file and once it finds a line which
starts with this string like
@STRING_A
then checks if 3 lines backwards there is another occurrence of a line
starting with the same string, like
@STRING_A
@STRING_A
and if it happens, to delete the occurrence 3 lines backward. I was thinking
about bash, but I do not know how to "go backwards" with it. So I am sure that
this is not possible with bash. I also thought about python, but then I should
store all information in memory in order to go backwards and then, for long
files it would be unfeasible.
What do you think? Is it possible to do it in bash or python?
Thanks
Answer: Funny that after all these hours nobody's yet given a solution to the problem
as actually phrased (as @John Machin points out in a comment) -- remove just
the leading marker (if followed by another such marker 3 lines down), not the
whole line containing it. It's not hard, of course -- here's a tiny mod as
needed of @truppo's fun solution, for example:
from itertools import izip, chain
f = "foo.txt"
for third, line in izip(chain(" ", open(f)), open(f)):
if third.startswith("@STRING_A") and line.startswith("@STRING_A"):
line = line[len("@STRING_A"):]
print line,
Of course, in real life, one would use an `iterator.tee` instead of reading
the file twice, have this code in a function, not repeat the marker constant
endlessly, &c;-).
|
Python Authentication with urllib2
Question: So I'm trying to download a file from a site called vsearch.cisco.com with
python
[python]
#Connects to the Cisco Server and Downloads files at the URL specified
import urllib2
#Define Useful Variables
url = 'http://vsearch.cisco.com'
username = 'xxxxxxxx'
password = 'xxxxxxxx'
realm = 'CEC'
# Begin Making connection
# Create a Handler -- Also could be where the error lies
handler = urllib2.HTTPDigestAuthHandler()
handler.add_password(realm,url,username,password)
# Create an Opener
opener = urllib2.build_opener(handler)
urllib2.install_opener(opener)
try:
urllib2.urlopen(url)
print f.read()
except urllib2.HTTPError, e:
print e.code
print e.header
[/python]
My error is ValueError: AbstractDigestAuthHandler doesn't know about basic
I've tried using Basic HTML Authorization handlers and even HTTPS handlers.
Nothing gives me access. This error is different from all the other errors
however. The other errors are simply 401 HTML errors
Any suggestions on how to do this?
Answer: A "password manager" might help:
mgr = urllib2.HTTPPasswordMgrWithDefaultRealm()
mgr.add_password(None, url, user, password)
urllib2.build_opener(urllib2.HTTPBasicAuthHandler(mgr),
urllib2.HTTPDigestAuthHandler(mgr))
|
How many numbers below N are coprimes to N?
Question: ## In short:
Given that **a** is coprime to **b** if **GCD(a,b) = 1** (where GCD stands for
[great common
divisor](http://en.wikipedia.org/wiki/Greatest%5Fcommon%5Fdivisor)), how many
positive integers below N are coprime to N?
Is there a clever way?
* * *
## Not necessary stuff
Here is the dumbest way:
def count_coprime(N):
counter = 0
for n in xrange(1,N):
if gcd(n,N) == 1:
counter += 1
return counter
It works, but it is slow, and dumb. I'd like to use a clever and faster
algorithm. I tried to use prime factors and divisors of N but I always get
something that doesn't work with larger N.
~~I think the algorithm should be able to count them without calculating all
of them like the dumbest algorithm does :P~~
## Edit
It seems I've found a working one:
def a_bit_more_clever_counter(N):
result = N - 1
factors = []
for factor, multiplicity in factorGenerator(N):
result -= N/factor - 1
for pf in factors:
if lcm(pf, factor) < N:
result += N/lcm(pf, factor) - 1
factors += [factor]
return result
where lcm is least common multiple. Does anyone have a better one?
## Note
I'm using python, I think code should be readable even to who doesn't know
python, if you find anything that is not clear just ask in the comments. I'm
interested in the algorithm and the math, the idea.
Answer: **[Edit]** One last thought, which (IMO) is important enough that I'll put it
at the beginning: if you're collecting a bunch of totients at once, you can
avoid a lot of redundant work. Don't bother starting from large numbers to
find their smaller factors -- instead, iterate over the smaller factors and
accumulate results for the larger numbers.
class Totient:
def __init__(self, n):
self.totients = [1 for i in range(n)]
for i in range(2, n):
if self.totients[i] == 1:
for j in range(i, n, i):
self.totients[j] *= i - 1
k = j / i
while k % i == 0:
self.totients[j] *= i
k /= i
def __call__(self, i):
return self.totients[i]
if __name__ == '__main__':
from itertools import imap
totient = Totient(10000)
print sum(imap(totient, range(10000)))
This takes just 8ms on my desktop.
* * *
The Wikipedia page on the [Euler totient
function](http://en.wikipedia.org/wiki/Euler_totient_function) has some nice
mathematical results.
![\\sum_{d\\mid
n}\\varphi\(d\)](http://chart.apis.google.com/chart?cht=tx&chl=%5Cnormalsize%5C%21%5Csum_%7Bd%5Cmid%20n%7D%5Cvarphi%28d%29)
counts the numbers coprime to and smaller than each divisor of
![n](http://chart.apis.google.com/chart?cht=tx&chl=%5Cnormalsize%5C%21n): this
has a trivial* mapping to counting the integers from
![1](http://chart.apis.google.com/chart?cht=tx&chl=%5Cnormalsize%5C%211) to
![n](http://chart.apis.google.com/chart?cht=tx&chl=%5Cnormalsize%5C%21n), so
the sum total is
![n](http://chart.apis.google.com/chart?cht=tx&chl=%5Cnormalsize%5C%21n).
_* by the second definition of_ [trivial](http://meshula.net/wordpress/?p=294)
This is perfect for an application of the [Möbius inversion
formula](http://en.wikipedia.org/wiki/M%C3%B6bius_inversion_formula), a clever
trick for inverting sums of this exact form.
![\\varphi\(n\)=\\sum_{d\\mid n}d\\cdot\\mu\\left\(\\frac
nd\\right\)](http://chart.apis.google.com/chart?cht=tx&chl=%5Cvarphi%28n%29%3D%5Csum_%7Bd%5Cmid%20n%7Dd%5Ccdot%5Cmu%5Cleft%28%5Cfrac%20nd%5Cright%29)
This leads naturally to the code
def totient(n):
if n == 1: return 1
return sum(d * mobius(n / d) for d in range(1, n+1) if n % d == 0)
def mobius(n):
result, i = 1, 2
while n >= i:
if n % i == 0:
n = n / i
if n % i == 0:
return 0
result = -result
i = i + 1
return result
There exist better implementations of the [Möbius
function](http://en.wikipedia.org/wiki/M%C3%B6bius_function), and it could be
memoized for speed, but this should be easy enough to follow.
The more obvious computation of the totient function is
![\\varphi\\left\(p_1^{k_1}\\dots
p_r^{k_r}\\right\)=\(p_1-1\)p_1^{k_1-1}\\dots\(p_r-1\)p_r^{k_r-1}p_1^{k_1}\\dots
p_r^{k_r}\\prod_{i=1}^r\\left\(1-\\frac1{p_r}\\right\)](http://chart.apis.google.com/chart?cht=tx&chl=%5Cvarphi%5Cleft%28p_1%5E%7Bk_1%7D%5Cdots%20p_r%5E%7Bk_r%7D%5Cright%29%3D%28p_1-1%29p_1%5E%7Bk_1-1%7D%5Cdots%28p_r-1%29p_r%5E%7Bk_r-1%7D%3Dp_1%5E%7Bk_1%7D%5Cdots%20p_r%5E%7Bk_r%7D%5Cprod_%7Bi%3D1%7D%5Er%5Cleft%281-%5Cfrac1%7Bp_r%7D%5Cright%29)
In other words, fully factor the number into unique primes and exponents, and
do a simple multiplication from there.
from operator import mul
def totient(n):
return int(reduce(mul, (1 - 1.0 / p for p in prime_factors(n)), n))
def primes_factors(n):
i = 2
while n >= i:
if n % i == 0:
yield i
n = n / i
while n % i == 0:
n = n / i
i = i + 1
Again, there exist better implementations of `prime_factors`, but this is
meant for easy reading.
* * *
`# helper functions`
from collections import defaultdict
from itertools import count
from operator import mul
def gcd(a, b):
while a != 0: a, b = b % a, a
return b
def lcm(a, b): return a * b / gcd(a, b)
primes_cache, prime_jumps = [], defaultdict(list)
def primes():
prime = 1
for i in count():
if i < len(primes_cache): prime = primes_cache[i]
else:
prime += 1
while prime in prime_jumps:
for skip in prime_jumps[prime]:
prime_jumps[prime + skip] += [skip]
del prime_jumps[prime]
prime += 1
prime_jumps[prime + prime] += [prime]
primes_cache.append(prime)
yield prime
def factorize(n):
for prime in primes():
if prime > n: return
exponent = 0
while n % prime == 0:
exponent, n = exponent + 1, n / prime
if exponent != 0:
yield prime, exponent
`# OP's first attempt`
def totient1(n):
counter = 0
for i in xrange(1, n):
if gcd(i, n) == 1:
counter += 1
return counter
`# OP's second attempt`
# I don't understand the algorithm, and just copying it yields inaccurate results
`# Möbius inversion`
def totient2(n):
if n == 1: return 1
return sum(d * mobius(n / d) for d in xrange(1, n+1) if n % d == 0)
mobius_cache = {}
def mobius(n):
result, stack = 1, [n]
for prime in primes():
if n in mobius_cache:
result = mobius_cache[n]
break
if n % prime == 0:
n /= prime
if n % prime == 0:
result = 0
break
stack.append(n)
if prime > n: break
for n in stack[::-1]:
mobius_cache[n] = result
result = -result
return -result
`# traditional formula`
def totient3(n):
return int(reduce(mul, (1 - 1.0 / p for p, exp in factorize(n)), n))
`# traditional formula, no division`
def totient4(n):
return reduce(mul, ((p-1) * p ** (exp-1) for p, exp in factorize(n)), 1)
Using this code to calculate the totients of all numbers from 1 to 9999 on my
desktop, averaging over 5 runs,
* `totient1` takes forever
* `totient2` takes 10s
* `totient3` takes 1.3s
* `totient4` takes 1.3s
|
How to replace a column using Python's built-in .csv writer module?
Question: I need to do a find and replace (specific to one column of URLs) in a huge
Excel .csv file. Since I'm in the beginning stages of trying to teach myself a
scripting language, I figured I'd try to implement the solution in python.
I'm having trouble with the "replace" part of the solution. I've read the
[official csv module documentation](http://docs.python.org/library/csv.html)
about how to use the writer, but there isn't really a clear enough example for
me (yes, I'm slow). So, now for the question: how does one iterate through the
rows of a csv file with a writer object?
p.s. apologies in advance for the clumsy code, I'm still learning :)
import csv
csvfile = open("PALTemplateData.csv")
csvout = open("PALTemplateDataOUT.csv")
dialect = csv.Sniffer().sniff(csvfile.read(1024))
csvfile.seek(0)
reader = csv.reader(csvfile, dialect)
writer = csv.writer(csvout, dialect)
total=0;
needchange=0;
changed = 0;
temp = ''
changeList = []
for row in reader:
total=total+1
temp = row[len(row)-1]
if '/?' in temp:
needchange=needchange+1;
changeList.append(row.index)
for row in writer: #this doesn't compile, hence the question
if row.index in changeList:
changed=changed+1
temp = row[len(row)-1]
temp.replace('/?', '?')
row[len(row)-1] = temp
writer.writerow(row)
print('Total URLs:', total)
print('Total URLs to change:', needchange)
print('Total URLs changed:', changed)
Answer: The reason you're getting an error is that the writer doesn't have data to
iterate over. You're supposed to give it the data - presumably, you'd have
some sort of list or generator that produces the rows to write out.
I'd suggest just combining the two loops, like so:
for row in reader:
row[-1] = row[-1].replace('/?', '?')
writer.writerow(row)
And with that, you don't even need `total`, `needchange`, and `changeList`.
(There are a bunch of optimizations in there that I unfortunately don't have
time to explain, but I'll see if I can edit that info in later)
|
Looping Fget with fsockopen in PHP 5.x
Question: I have a Python Server finally working and responding to multiple command's
with the output's, however I'm now having problem's with PHP receiving the
full output. I have tried commands such as fgets, fread, the only command that
seems to work is "fgets".
However this only recieve's on line of data, I then created a while statement
shown bellow:
while (!feof($handle)) {
$buffer = fgets($handle, 4096);
echo $buffer;
}
However it seems the Python server is not sending a Feof at the end of the
output so the php page times out and does not display anything. Like I said
above, just running echo fgets($handle), work's fine, and output's one line,
running the command again under neither will display the next line e.t.c
I have attached the important part of my Python Script bellow:
s = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
s.bind(("", port))
s.listen(5)
print "OK."
print " Listening on port:", port
import subprocess
while 1:
con, addr = s.accept()
while True:
datagram = con.recv(1024)
if not datagram:
break
print "Rx Cmd:", datagram
print "Launch:", datagram
process = subprocess.Popen(datagram+" &", shell=True, stdout=subprocess.PIPE)
stdout, stderr = process.communicate()
con.send(stdout)
con.close()
s.close()
I have also attached the full PHP script:
<?php
$handle = fsockopen("tcp://xxx.xxx.xxx.xxx",12345);
fwrite($handle,"ls");
echo fgets($handle);
fclose($handle);
?>
Thanks, Ashley
Answer: I believe you need to fix your server code a bit. I have removed the inner
while loop. The problem with your code was that the server never closed the
connection, so `feof` never returned true.
I also removed the `+ " &"` bit. To get the output, you need to wait until the
process ends anyway. And I am not sure how the shell would handle the `&` in
this case.
s = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
s.bind(("", port))
s.listen(5)
print "OK."
print " Listening on port:", port
import subprocess
try:
while 1:
con, addr = s.accept()
try:
datagram = con.recv(1024)
if not datagram:
continue
print "Rx Cmd:", datagram
print "Launch:", datagram
process = subprocess.Popen(datagram, shell=True, stdout=subprocess.PIPE)
stdout, stderr = process.communicate()
con.send(stdout)
finally:
print "closing connection"
con.close()
except KeyboardInterrupt:
pass
finally:
print "closing socket"
s.close()
BTW, you need to use the while-loop in your php script. `fgets` returns until
a single line only.
|
How do I set up a basic website with registration in Python on Dreamhost?
Question: I need to write a basic website on Dreamhost. It needs to be done in Python. I
discovered Dreamhost permits me to write .py files, and read them.
### Example:
#!/usr/bin/python
print "Content-type: text/html\n\n"
print "hello world"
So now I am looking for a basic framework, or a set of files that has already
programmed the whole registration to be able to kick-off the project in a
simple way. By registration I mean the files to register a new account, log
in, check the email (sending a mail), and edit the user information. All this
possibly using MySQL.
Answer: Let me share my own experience with django. My prerequisits:
* average knowledge of python
* very weak idea of how web works (no js skills, just a bit of css)
* my day job is filled with coding in C and I just wanted to try something different, so there certainly was a passion to learn (I think this is the most important one)
Why I've chosen django:
* I've already knew bits and pieces of python
* django has excelent documentation, including tutorial, which explained everything in very clear and simple manner
It is worth to read complete [manual](http://docs.djangoproject.com/en/dev/)
first (it took me two or three weekends. I remember I could not
remember/understand everything at first pass, but it helped me to learn where
the information can be found when needed. There is also another source of
documentaion called [djangobook](http://www.djangobook.com/ "djangobook").
Djangobook contains same information as manual, but things are explained more
in detail. It's worth to read it also, it helps to catch up with MVC concept,
if you have not tried that before.
And finally to answer your question best: there are already also
[OpenId](http://openid.net) modules ready for you. I'm considering to use
[django-authopenid](http://bitbucket.org/benoitc/django-authopenid/wiki/Home)
for my new project. It supports OpenId, while providing fallback to locally
managed users.
There is certain learning curve if you are going learn django. The more you
know about the web and python the steeper the curve is. I had to also learn
bits and pieces of javascript and it took me also some time. If you are able
to spend full time learning django, then you can expect you'll be able to
deliver first results within 4-6 weeks. It took me 6 months, since I was doing
my django studies in free time.
|
Formatting cells in Excel with Python
Question: How do I format cells in Excel with python?
In particular I need to change the font of several subsequent rows to be
regular instead of bold.
Thnak you,
Alex
Answer: Using [xlwt](http://pypi.python.org/pypi/xlwt):
from xlwt import *
font0 = Font()
font0.bold = False
style0 = XFStyle()
style0.font = font0
wb = Workbook()
ws0 = wb.add_sheet('0')
ws0.write(0, 0, 'myNormalText', style0)
font1 = Font()
font1.bold = True
style1 = XFStyle()
style1.font = font1
ws0.write(0, 1, 'myBoldText', style1)
wb.save('format.xls')
|
Flash Characters on Screen in Linux
Question: I have a XFCE 4.6 on kernel 2.6. Is there a quick and easy way to flash a
message on the screen for a few seconds?
My Thinkpad T60 has 3 volume buttons (up, down, mute). When I pressed the
buttons, I would like to flash the volume on the screen for a second on
screen. Can it be done with Python?
Answer: [notification-daemon-
xfce](http://goodies.xfce.org/projects/applications/notification-daemon-xfce)
allows [libnotify](http://www.galago-project.org/) clients to show brief
messages in XFCE. libnotify has [Python bindings](http://www.galago-
project.org/files/releases/source/notify-python/) available.
As an untested example,
import pynotify
import sys
pynotify.init(sys.argv[0])
notification = pynotify.Notification("Title", "body", "dialog-info")
notification.set_urgency(pynotify.URGENCY_NORMAL)
notification.set_timeout(pynotify.EXPIRES_DEFAULT)
notification.show()
|
Reading binary file in Python and looping over each byte
Question: In Python, how do I read in a binary file and loop over each byte of that
file?
Answer:
f = open("myfile", "rb")
try:
byte = f.read(1)
while byte != "":
# Do stuff with byte.
byte = f.read(1)
finally:
f.close()
By suggestion of chrispy:
with open("myfile", "rb") as f:
byte = f.read(1)
while byte != "":
# Do stuff with byte.
byte = f.read(1)
Note that the with statement is not available in versions of Python below 2.5.
To use it in v 2.5 you'll need to import it:
from __future__ import with_statement
In 2.6 this is not needed.
In Python 3, it's a bit different. We will no longer get raw characters from
the stream in byte mode but byte objects, thus we need to alter the condition:
with open("myfile", "rb") as f:
byte = f.read(1)
while byte != b"":
# Do stuff with byte.
byte = f.read(1)
Or as benhoyt says, skip the not equal and take advantage of the fact that
`b""` evaluates to false. This makes the code compatible between 2.6 and 3.x
without any changes. It would also save you from changing the condition if you
go from byte mode to text or the reverse.
with open("myfile", "rb") as f:
byte = f.read(1)
while byte:
# Do stuff with byte.
byte = f.read(1)
|
In Python, how do I easily generate an image file from some source data?
Question: I have some some data that I would like to visualize. Each byte of the source
data roughly corresponds to a pixel value of the image.
What is the easiest way to generate an image file (bitmap?) using Python?
Answer: You can create images with a list of pixel values using
[Pillow](http://python-pillow.github.io/):
from PIL import Image
img = Image.new('RGB', (width, height))
img.putdata(my_list)
img.save('image.png')
|
wx.Panel scales to fit entire parent Frame despite giving it a size
Question: Hi I am newbie to wxpython I am trying to have a Frame and within that a small
panel area which I am coloring blue. However no matter what I do the wx.Panel
using the size attribute , the single panel snaps to the size of its parent
frame. If I add another panel (pane2 in code below) both panes are drawn in
the correct size.
I know I can control these panels using sizers . But I was trying to
understand why the wx.Panel object behaves the way it does when its all alone.
Here is the code:
import wx
class PlateGui(wx.Frame):
def __init__(self, *args , **kwds):
self.frame = wx.Frame.__init__(self,*args, **kwds)
print "Made frame"
if __name__ == "__main__":
an_app = wx.PySimpleApp()
aframe = PlateGui(parent=None,id=-1,title="Test Frame",size=(300, 300))
pane = wx.Panel(parent=aframe,size=(100,100),style=wx.RAISED_BORDER)
pane.SetBackgroundColour(wx.Colour(0,0,255))
# pane2 = wx.Panel(parent=aframe,size=(200,100),style=wx.RAISED_BORDER)
# Commenting out the second pane makes the first pane fit
# entire frame regardless of size specified
aframe.Show()
an_app.MainLoop()
Answer: By default, wx.Frame has a sizer that expands its child to fill the frame.
Create your own sizer, add the panel to it (without specifying expand flags)
and set that as the frame's sizer.
import wx
app = wx.PySimpleApp()
frame = wx.Frame(None, -1, 'Test')
sizer = wx.BoxSizer(wx.VERTICAL)
panel = wx.Panel(frame, -1, size=(100,100), style=wx.BORDER_RAISED)
sizer.Add(panel)
frame.SetSizer(sizer)
frame.Show()
app.MainLoop()
|
Django unit testing with date/time-based objects
Question: Suppose I have the following `Event` model:
from django.db import models
import datetime
class Event(models.Model):
date_start = models.DateField()
date_end = models.DateField()
def is_over(self):
return datetime.date.today() > self.date_end
I want to test `Event.is_over()` by creating an Event that ends in the future
(today + 1 or something), and stubbing the date and time so the system thinks
we've reached that future date.
I'd like to be able to stub ALL system time objects as far as python is
concerned. This includes `datetime.date.today()`, `datetime.datetime.now()`,
and any other standard date/time objects.
What's the standard way to do this?
Answer: **EDIT** : Since my answer is the accepted answer here I'm updating it to let
everyone know a better way has been created in the meantime, the freezegun
library: <https://pypi.python.org/pypi/freezegun>. I use this in all my
projects when I want to influence time in tests. Have a look at it.
**Original answer:**
Replacing internal stuff like this is always dangerous because it can have
nasty side effects. So what you indeed want, is to have the monkey patching be
as local as possible.
We use Michael Foord's excellent mock library:
<http://www.voidspace.org.uk/python/mock/> that has a `@patch` decorator which
patches certain functionality, but the monkey patch only lives in the scope of
the testing function, and everything is automatically restored after the
function runs out of its scope.
The only problem is that the internal `datetime` module is implemented in C,
so by default you won't be able to monkey patch it. We fixed this by making
our own simple implementation which _can_ be mocked.
The total solution is something like this (the example is a validator function
used within a Django project to validate that a date is in the future). Mind
you I took this from a project but took out the non-important stuff, so things
may not actually work when copy-pasting this, but you get the idea, I hope :)
First we define our own very simple implementation of `datetime.date.today` in
a file called `utils/date.py`:
import datetime
def today():
return datetime.date.today()
Then we create the unittest for this validator in `tests.py`:
import datetime
import mock
from unittest2 import TestCase
from django.core.exceptions import ValidationError
from .. import validators
class ValidationTests(TestCase):
@mock.patch('utils.date.today')
def test_validate_future_date(self, today_mock):
# Pin python's today to returning the same date
# always so we can actually keep on unit testing in the future :)
today_mock.return_value = datetime.date(2010, 1, 1)
# A future date should work
validators.validate_future_date(datetime.date(2010, 1, 2))
# The mocked today's date should fail
with self.assertRaises(ValidationError) as e:
validators.validate_future_date(datetime.date(2010, 1, 1))
self.assertEquals([u'Date should be in the future.'], e.exception.messages)
# Date in the past should also fail
with self.assertRaises(ValidationError) as e:
validators.validate_future_date(datetime.date(2009, 12, 31))
self.assertEquals([u'Date should be in the future.'], e.exception.messages)
The final implementation looks like this:
from django.utils.translation import ugettext_lazy as _
from django.core.exceptions import ValidationError
from utils import date
def validate_future_date(value):
if value <= date.today():
raise ValidationError(_('Date should be in the future.'))
Hope this helps
|
MS Access library for python
Question: Is there a library for using MS Access database in python? The win32 module is
not as easy as the MySQL library. Is there a simpler way to use MS Access with
Python?
Answer: Depending on what you want to do,
[pyodbc](https://github.com/mkleehammer/pyodbc) might be what you are looking
for.
import pyodbc
db_file = r'''C:\x.mdb'''
user = 'admin'
password = ''
odbc_conn_str = 'DRIVER={Microsoft Access Driver (*.mdb)};DBQ=%s;UID=%s;PWD=%s' %\
(db_file, user, password)
# Or, for newer versions of the Access drivers:
odbc_conn_str = 'DRIVER={Microsoft Access Driver (*.mdb, *.accdb)};DBQ=%s;UID=%s;PWD=%s' %\
(db_file, user, password)
conn = pyodbc.connect(odbc_conn_str)
|
wxPython SplitterWindow does not expand within a Panel
Question: I'm trying a simple layout and the panel divided by a SplitterWindow doesn't
expand to fill the whole area, what I want is this:
[button] <= (fixed size)
---------
TEXT AREA }
~~~~~~~~~ <= (this is the splitter) } this is a panel
TEXT AREA }
The actual code is:
import wx
app = wx.App()
frame = wx.Frame(None, wx.ID_ANY, "Register Translator")
parseButton = wx.Button(frame, label="Parse")
panel = wx.Panel(frame)
panel.SetBackgroundColour("BLUE")
splitter = wx.SplitterWindow(panel)
inputArea = wx.TextCtrl(splitter, style=wx.TE_MULTILINE)
outputArea = wx.TextCtrl(splitter, style=wx.TE_MULTILINE)
splitter.SplitHorizontally(inputArea, outputArea)
sizer=wx.BoxSizer(wx.VERTICAL)
sizer.Add(parseButton, 0, wx.ALIGN_CENTER)
sizer.Add(panel, 1, wx.EXPAND | wx.ALL)
frame.SetSizerAndFit(sizer)
frame.SetAutoLayout(1)
frame.Show(True)
app.MainLoop()
I set the panel color different, and it's actually using the whole area, so
the problem is just the SplitterWindow within the Panel, not within the
BoxSizer.
Any ideas about why it isn't working? Thanks!
Answer: The Panel is probably expanding but the ScrolledWindow within the Panel is
not, because you aren't using a sizer for the panel, only the frame.
You could also try just having the SplitterWindow be a child of the frame,
without the panel.
|
Importing database data into Joomla
Question: How to import data from a database to Joomla CMS?
I have a database with lots of data I want to use in my new website. An ideal
solution for me would be a Python/Perl/PHP API that would know how to do
Joomla' basic routines:
1. adding/removing a section/category/material/menu/module;
2. changing properties of existing entities
Answer: You could try the following extensions:
1. [Bulk Import](http://extensions.joomla.org/extensions/migration-&-conversion/data-import-&-export/7243/details)
2. [CSV Import](http://extensions.joomla.org/extensions/migration-&-conversion/extensions-migration/4247/details)
If that doesn't work for you, maybe take a look at the [Joomla
API](http://api.joomla.org/)
|
formencode invalid return type
Question: if an exception occurs in form encode then what will be the return type??
suppose
if(request.POST):
formvalidate = ValidationRule()
try:
new = formvalidate.to_python(request.POST)
data = Users1( n_date = new['n_date'], heading = new['heading'],
desc = new['desc'], link = new['link'],
module_name = new['module_name'] )
session.add(data)
session.commit()
except formencode.Invalid, e:
errors = e
how we can find the field wise error
Answer: I assume you are using formencode(<http://formencode.org>)
you can use unpack_errors to get per field error e.g.
import formencode
from formencode import validators
class UserForm(formencode.Schema):
first_name = validators.String(not_empty=True)
last_name = validators.String(not_empty=True)
form = UserForm()
try:
form.to_python({})
except formencode.Invalid,e:
print e.unpack_errors()
it will print a dict of errors per field.
you can use formencode.htmlfill.render to render all errors, in different
ways, read <http://formencode.org/htmlfill.html#errors>
|
KenKen puzzle addends: REDUX A (corrected) non-recursive algorithm
Question: This question relates to those parts of the KenKen Latin Square puzzles which
ask you to find all possible combinations of ncells numbers with values x such
that 1 <= x <= maxval and x(1) + ... + x(ncells) = targetsum. Having tested
several of the more promising answers, I'm going to award the answer-prize to
Lennart Regebro, because:
1. his routine is as fast as mine (+-5%), and
2. he pointed out that my original routine had a bug somewhere, which led me to see what it was really trying to do. Thanks, Lennart.
chrispy contributed an algorithm that seems equivalent to Lennart's, but 5 hrs
later, sooo, first to the wire gets it.
A remark: Alex Martelli's bare-bones recursive algorithm is an example of
making every possible combination and throwing them all at a sieve and seeing
which go through the holes. This approach takes 20+ times longer than
Lennart's or mine. (Jack up the input to max_val = 100, n_cells = 5,
target_sum = 250 and on my box it's 18 secs vs. 8+ mins.) Moral: Not
generating every possible combination is good.
Another remark: Lennart's and my routines generate **the same answers in the
same order**. Are they in fact the same algorithm seen from different angles?
I don't know.
Something occurs to me. If you sort the answers, starting, say, with
(8,8,2,1,1) and ending with (4,4,4,4,4) (what you get with max_val=8,
n_cells=5, target_sum=20), the series forms kind of a "slowest descent", with
the first ones being "hot" and the last one being "cold" and the greatest
possible number of stages in between. Is this related to "informational
entropy"? What's the proper metric for looking at it? Is there an algorithm
that producs the combinations in descending (or ascending) order of heat?
(This one doesn't, as far as I can see, although it's close over short
stretches, looking at normalized std. dev.)
Here's the Python routine:
#!/usr/bin/env python
#filename: makeAddCombos.07.py -- stripped for StackOverflow
def initialize_combo( max_val, n_cells, target_sum):
"""returns combo
Starting from left, fills combo to max_val or an intermediate value from 1 up.
E.g.: Given max_val = 5, n_cells=4, target_sum = 11, creates [5,4,1,1].
"""
combo = []
#Put 1 in each cell.
combo += [1] * n_cells
need = target_sum - sum(combo)
#Fill as many cells as possible to max_val.
n_full_cells = need //(max_val - 1)
top_up = max_val - 1
for i in range( n_full_cells): combo[i] += top_up
need = target_sum - sum(combo)
# Then add the rest to next item.
if need > 0:
combo[n_full_cells] += need
return combo
#def initialize_combo()
def scrunch_left( combo):
"""returns (new_combo,done)
done Boolean; if True, ignore new_combo, all done;
if Falso, new_combo is valid.
Starts a new combo list. Scanning from right to left, looks for first
element at least 2 greater than right-end element.
If one is found, decrements it, then scrunches all available counts on its
right up against its right-hand side. Returns the modified combo.
If none found, (that is, either no step or single step of 1), process
done.
"""
new_combo = []
right_end = combo[-1]
length = len(combo)
c_range = range(length-1, -1, -1)
found_step_gt_1 = False
for index in c_range:
value = combo[index]
if (value - right_end) > 1:
found_step_gt_1 = True
break
if not found_step_gt_1:
return ( new_combo,True)
if index > 0:
new_combo += combo[:index]
ceil = combo[index] - 1
new_combo += [ceil]
new_combo += [1] * ((length - 1) - index)
need = sum(combo[index:]) - sum(new_combo[index:])
fill_height = ceil - 1
ndivf = need // fill_height
nmodf = need % fill_height
if ndivf > 0:
for j in range(index + 1, index + ndivf + 1):
new_combo[j] += fill_height
if nmodf > 0:
new_combo[index + ndivf + 1] += nmodf
return (new_combo, False)
#def scrunch_left()
def make_combos_n_cells_ge_two( combos, max_val, n_cells, target_sum):
"""
Build combos, list of tuples of 2 or more addends.
"""
combo = initialize_combo( max_val, n_cells, target_sum)
combos.append( tuple( combo))
while True:
(combo, done) = scrunch_left( combo)
if done:
break
else:
combos.append( tuple( combo))
return combos
#def make_combos_n_cells_ge_two()
if __name__ == '__main__':
combos = []
max_val = 8
n_cells = 5
target_sum = 20
if n_cells == 1: combos.append( (target_sum,))
else:
combos = make_combos_n_cells_ge_two( combos, max_val, n_cells, target_sum)
import pprint
pprint.pprint( combos)
Answer: Your algorithm seems pretty good at first blush, and I don't think OO or
another language would improve the code. I can't say if recursion would have
helped but I admire the non-recursive approach. I bet it was harder to get
working and it's harder to read but it likely is more efficient and it's
definitely quite clever. To be honest I didn't analyze the algorithm in detail
but it certainly looks like something that took a long while to get working
correctly. I bet there were lots of off-by-1 errors and weird edge cases you
had to think through, eh?
Given all that, basically all I tried to do was pretty up your code as best I
could by replacing the numerous C-isms with more idiomatic Python-isms. Often
times what requires a loop in C can be done in one line in Python. Also I
tried to rename things to follow Python naming conventions better and cleaned
up the comments a bit. Hope I don't offend you with any of my changes. You can
take what you want and leave the rest. :-)
Here are the notes I took as I worked:
* Changed the code that initializes `tmp` to a bunch of 1's to the more idiomatic `tmp = [1] * n_cells`.
* Changed `for` loop that sums up `tmp_sum` to idiomatic `sum(tmp)`.
* Then replaced all the loops with a `tmp = <list> + <list>` one-liner.
* Moved `raise doneException` to `init_tmp_new_ceiling` and got rid of the `succeeded` flag.
* The check in `init_tmp_new_ceiling` actually seems unnecessary. Removing it, the only `raise`s left were in `make_combos_n_cells`, so I just changed those to regular returns and dropped `doneException` entirely.
* Normalized mix of 4 spaces and 8 spaces for indentation.
* Removed unnecessary parentheses around your `if` conditions.
* `tmp[p2] - tmp[p1] == 0` is the same thing as `tmp[p2] == tmp[p1]`.
* Changed `while True: if new_ceiling_flag: break` to `while not new_ceiling_flag`.
* You don't need to initialize variables to 0 at the top of your functions.
* Removed `combos` list and changed function to `yield` its tuples as they are generated.
* Renamed `tmp` to `combo`.
* Renamed `new_ceiling_flag` to `ceiling_changed`.
And here's the code for your perusal:
def initial_combo(ceiling=5, target_sum=13, num_cells=4):
"""
Returns a list of possible addends, probably to be modified further.
Starts a new combo list, then, starting from left, fills items to ceiling
or intermediate between 1 and ceiling or just 1. E.g.:
Given ceiling = 5, target_sum = 13, num_cells = 4: creates [5,5,2,1].
"""
num_full_cells = (target_sum - num_cells) // (ceiling - 1)
combo = [ceiling] * num_full_cells \
+ [1] * (num_cells - num_full_cells)
if num_cells > num_full_cells:
combo[num_full_cells] += target_sum - sum(combo)
return combo
def all_combos(ceiling, target_sum, num_cells):
# p0 points at the rightmost item and moves left under some conditions
# p1 starts out at rightmost items and steps left
# p2 starts out immediately to the left of p1 and steps left as p1 does
# So, combo[p2] and combo[p1] always point at a pair of adjacent items.
# d combo[p2] - combo[p1]; immediate difference
# cd combo[p2] - combo[p0]; cumulative difference
# The ceiling decreases by 1 each iteration.
while True:
combo = initial_combo(ceiling, target_sum, num_cells)
yield tuple(combo)
ceiling_changed = False
# Generate all of the remaining combos with this ceiling.
while not ceiling_changed:
p2, p1, p0 = -2, -1, -1
while combo[p2] == combo[p1] and abs(p2) <= num_cells:
# 3,3,3,3
if abs(p2) == num_cells:
return
p2 -= 1
p1 -= 1
p0 -= 1
cd = 0
# slide_ptrs_left loop
while abs(p2) <= num_cells:
d = combo[p2] - combo[p1]
cd += d
# 5,5,3,3 or 5,5,4,3
if cd > 1:
if abs(p2) < num_cells:
# 5,5,3,3 --> 5,4,4,3
if d > 1:
combo[p2] -= 1
combo[p1] += 1
# d == 1; 5,5,4,3 --> 5,4,4,4
else:
combo[p2] -= 1
combo[p0] += 1
yield tuple(combo)
# abs(p2) == num_cells; 5,4,4,3
else:
ceiling -= 1
ceiling_changed = True
# Resume at make_combo_same_ceiling while
# and follow branch.
break
# 4,3,3,3 or 4,4,3,3
elif cd == 1:
if abs(p2) == num_cells:
return
p1 -= 1
p2 -= 1
if __name__ == '__main__':
print list(all_combos(ceiling=6, target_sum=12, num_cells=4))
|
Adding Cookie to SOAPpy Request
Question: I'm trying to send a SOAP request using SOAPpy as the client. I've found some
documentation stating how to add a cookie by extending SOAPpy.HTTPTransport,
but I can't seem to get it to work.
I tried to use the example
[here](http://code.activestate.com/recipes/444758/), but the server I'm trying
to send the request to started throwing 415 errors, so I'm trying to
accomplish this without using ClientCookie, or by figuring out why the server
is throwing 415's when I do use it. I suspect it might be because ClientCookie
uses urllib2 & http/1.1, whereas SOAPpy uses urllib & http/1.0
Does someone know how to make ClientCookie use http/1.0, if that is even the
problem, or a way to add a cookie to the SOAPpy headers without using
ClientCookie? If tried this code using other services, it only seems to throw
errors when sending requests to Microsoft servers.
I'm still finding my footing with python, so it could just be me doing
something dumb.
import sys, os, string
from SOAPpy import WSDL,HTTPTransport,Config,SOAPAddress,Types
import ClientCookie
Config.cookieJar = ClientCookie.MozillaCookieJar()
class CookieTransport(HTTPTransport):
def call(self, addr, data, namespace, soapaction = None, encoding = None,
http_proxy = None, config = Config):
if not isinstance(addr, SOAPAddress):
addr = SOAPAddress(addr, config)
cookie_cutter = ClientCookie.HTTPCookieProcessor(config.cookieJar)
hh = ClientCookie.HTTPHandler()
hh.set_http_debuglevel(1)
# TODO proxy support
opener = ClientCookie.build_opener(cookie_cutter, hh)
t = 'text/xml';
if encoding != None:
t += '; charset="%s"' % encoding
opener.addheaders = [("Content-Type", t),
("Cookie", "Username=foobar"), # ClientCookie should handle
("SOAPAction" , "%s" % (soapaction))]
response = opener.open(addr.proto + "://" + addr.host + addr.path, data)
data = response.read()
# get the new namespace
if namespace is None:
new_ns = None
else:
new_ns = self.getNS(namespace, data)
print '\n' * 4 , '-'*50
# return response payload
return data, new_ns
url = 'http://www.authorstream.com/Services/Test.asmx?WSDL'
proxy = WSDL.Proxy(url, transport=CookieTransport)
print proxy.GetList()
Answer: Error 415 is because of incorrect content-type header.
Install httpfox for firefox or whatever tool (wireshark, Charles or Fiddler)
to track what headers are you sending. Try Content-Type: application/xml.
...
t = 'application/xml';
if encoding != None:
t += '; charset="%s"' % encoding
...
If you trying to send file to the web server use Content-
Type:application/x-www-form-urlencoded
|
In Python, Using pyodbc, How Do You Perform Transactions?
Question: I have a username which I must change in numerous (up to ~25) tables. (Yeah, I
know.) An atomic transaction seems to be the way to go for this sort of thing.
However, I do not know how to do this with pyodbc. I've seen various tutorials
on atomic transactions before, but have never used them.
The setup: Windows platform, Python 2.6, pyodbc, Microsoft SQL 2005. I've used
pyodbc for single SQL statements, but no compound statements or transactions.
Best practices for SQL seem to suggest that creating a stored procedure is
excellent for this. My fears about doing a stored procedure are as follows, in
order of increasing importance: 1) I have never written a stored procedure. 2)
I heard that pyodbc does not return results from stored procedures as of yet.
3) This is most definitely Not My Database. It's vendor-supplied, vendor-
updated, and so forth.
So, what's the best way to go about this?
Answer: By its
[documentation](http://code.google.com/p/pyodbc/wiki/FAQs#Connecting%5Ffails%5Fwith%5Fan%5Ferror%5Fabout%5FSQL%5FATTR%5FAUTOCOMMIT),
pyodbc does support transactions, but only if the odbc driver support it.
Furthermore, as pyodbc is compliant with [PEP
249](http://www.python.org/dev/peps/pep-0249/), data is stored only when a
manual commit is done.
However, it also support autocommit feature, and in that case you cannot have
any transaction.
You should check the connection, when it is performed
cnxn = pyodbc.connect(cstring, autocommit=True)
or explicitely turn off the autocommit mode with
cnxn.autocommit = False
Note: you can get more information on the autocommit mode of pyodbc on its
[wiki](http://code.google.com/p/pyodbc/wiki/Features)
Once autocommit is turned off, then you have to explicitely commit() the
transaction, or rollback() the entire transaction.
|
Translating Perl to Python
Question: I found this Perl script while [migrating my SQLite database to
mysql](http://stackoverflow.com/questions/18671/quick-easy-way-to-migrate-
sqlite3-to-mysql/25860)
I was wondering (since I don't know Perl) how could one rewrite this in
Python?
Bonus points for the shortest (code) answer :)
**edit** : sorry I meant shortest code, not strictly shortest answer
#! /usr/bin/perl
while ($line = <>){
if (($line !~ /BEGIN TRANSACTION/) && ($line !~ /COMMIT/) && ($line !~ /sqlite_sequence/) && ($line !~ /CREATE UNIQUE INDEX/)){
if ($line =~ /CREATE TABLE \"([a-z_]*)\"(.*)/){
$name = $1;
$sub = $2;
$sub =~ s/\"//g; #"
$line = "DROP TABLE IF EXISTS $name;\nCREATE TABLE IF NOT EXISTS $name$sub\n";
}
elsif ($line =~ /INSERT INTO \"([a-z_]*)\"(.*)/){
$line = "INSERT INTO $1$2\n";
$line =~ s/\"/\\\"/g; #"
$line =~ s/\"/\'/g; #"
}else{
$line =~ s/\'\'/\\\'/g; #'
}
$line =~ s/([^\\'])\'t\'(.)/$1THIS_IS_TRUE$2/g; #'
$line =~ s/THIS_IS_TRUE/1/g;
$line =~ s/([^\\'])\'f\'(.)/$1THIS_IS_FALSE$2/g; #'
$line =~ s/THIS_IS_FALSE/0/g;
$line =~ s/AUTOINCREMENT/AUTO_INCREMENT/g;
print $line;
}
}
Some additional code was necessary to successfully migrate the sqlite database
(handles one line Create table statements, foreign keys, fixes a bug in the
original program that converted empty fields `''` to `\'`.
I [posted the code on the migrating my SQLite database to mysql
Question](http://stackoverflow.com/questions/18671/quick-easy-way-to-migrate-
sqlite3-to-mysql/1067365#1067365)
Answer: Here's a pretty literal translation with just the minimum of obvious style
changes (putting all code into a function, using string rather than re
operations where possible).
import re, fileinput
def main():
for line in fileinput.input():
process = False
for nope in ('BEGIN TRANSACTION','COMMIT',
'sqlite_sequence','CREATE UNIQUE INDEX'):
if nope in line: break
else:
process = True
if not process: continue
m = re.search('CREATE TABLE "([a-z_]*)"(.*)', line)
if m:
name, sub = m.groups()
line = '''DROP TABLE IF EXISTS %(name)s;
CREATE TABLE IF NOT EXISTS %(name)s%(sub)s
'''
line = line % dict(name=name, sub=sub)
else:
m = re.search('INSERT INTO "([a-z_]*)"(.*)', line)
if m:
line = 'INSERT INTO %s%s\n' % m.groups()
line = line.replace('"', r'\"')
line = line.replace('"', "'")
line = re.sub(r"([^'])'t'(.)", r"\1THIS_IS_TRUE\2", line)
line = line.replace('THIS_IS_TRUE', '1')
line = re.sub(r"([^'])'f'(.)", r"\1THIS_IS_FALSE\2", line)
line = line.replace('THIS_IS_FALSE', '0')
line = line.replace('AUTOINCREMENT', 'AUTO_INCREMENT')
print line,
main()
|
Get rid of toplevel tk panewindow while usong tkMessageBox
Question: [link text](http://stackoverflow.com/questions/1052420/tkkinter-message-box)
When I do :
tkMessageBox.askquestion(title="Symbol Display",message="Is the symbol visible on the console")
along with Symbol Display window tk window is also coming.
If i press "Yes"...the child window return yes,whereas tk window remains
there.
Whenever I am tryng to close tk window, End Program - tk comes. on pushing
"End Now" button "pythonw.exe" window comes asking to send error report or
not.
Why is it so ? How can I avoid tk window from popping out without affecting my
script execution ???
Answer: The _trick_ is to invoke withdraw on the Tk root top-level:
>>> import tkMessageBox, Tkinter
>>> Tkinter.Tk().withdraw()
>>> tkMessageBox.askquestion(
... title="Symbol Display",
... message="Is the symbol visible on the console")
|
Using SimpleXMLTreeBuilder in elementtree
Question: I have been developing an application with django and elementtree and while
deploying it to the production server i have found out it is running python
2.4. I have been able to bundle elementtree but now i am getting the error:
"No module named expat; use SimpleXMLTreeBuilder instead"
Unfortunately i cannot upgrade python so im stuck with what i got. How do i
use SimpleXMLTreeBuilder as the parser and/or will i need to rewrite code?
Answer: If you have third party module that wants to use ElementTree (and
XMLTreeBuilder by dependency) you can change ElementTree's XMLTreeBuilder
definition to the one provided by SimpleXMLTreeBuilder like so:
from xml.etree import ElementTree # part of python distribution
from elementtree import SimpleXMLTreeBuilder # part of your codebase
ElementTree.XMLTreeBuilder = SimpleXMLTreeBuilder.TreeBuilder
Now ElementTree will always use the SimpleXMLTreeBuilder whenever it's called.
See also: <http://groups.google.com/group/google-
appengine/browse_thread/thread/b7399a91c9525c97>
|
Implementing a custom Python authentication handler
Question: The answer to a [previous
question](http://stackoverflow.com/questions/1080179/handling-authentication-
and-proxy-servers-with-httplib2) showed that Nexus implement a [custom
authentication helper](http://svn.sonatype.org/nexus/tags/nexus-1.3.4/nexus-
clients/nexus-rest-client-
java/src/main/java/org/sonatype/nexus/client/rest/HttpNxBasicHelper.java)
called "NxBASIC".
How do I begin to implement a handler in python?
* * *
Update:
Implementing the handler per Alex's suggestion looks to be the right approach,
but fails trying to extract the scheme and realm from the authreq. The
returned value for authreq is:
str: NxBASIC realm="Sonatype Nexus Repository Manager API""
AbstractBasicAuthHandler.rx.search(authreq) is only returning a single tuple:
tuple: ('NxBASIC', '"', 'Sonatype Nexus Repository Manager API')
so scheme,realm = mo.groups() fails. From my limited regex knowledge it looks
like the standard regex from AbstractBasicAuthHandler should match scheme and
realm, but it seems not to.
The regex is:
rx = re.compile('(?:.*,)*[ \t]*([^ \t]+)[ \t]+'
'realm=(["\'])(.*?)\\2', re.I)
* * *
Update 2: From inspection of AbstractBasicAuthHandler, the default processing
is to do:
scheme, quote, realm = mo.groups()
Changing to this works. I now just need to set the password against the
correct realm. Thanks Alex!
Answer: If, as described, name and description are the only differences between this
"NxBasic" and good old "Basic", then you could essentially copy-paste-edit
some code from urllib2.py (which unfortunately doesn't expose the scheme name
as easily overridable in itself), as follows (see
[urllib2.py](http://svn.python.org/view/python/trunk/Lib/urllib2.py?revision=72880&view=markup)'s
online sources):
import urllib2
class HTTPNxBasicAuthHandler(urllib2.HTTPBasicAuthHandler):
def http_error_auth_reqed(self, authreq, host, req, headers):
# host may be an authority (without userinfo) or a URL with an
# authority
# XXX could be multiple headers
authreq = headers.get(authreq, None)
if authreq:
mo = AbstractBasicAuthHandler.rx.search(authreq)
if mo:
scheme, realm = mo.groups()
if scheme.lower() == 'nxbasic':
return self.retry_http_basic_auth(host, req, realm)
def retry_http_basic_auth(self, host, req, realm):
user, pw = self.passwd.find_user_password(realm, host)
if pw is not None:
raw = "%s:%s" % (user, pw)
auth = 'NxBasic %s' % base64.b64encode(raw).strip()
if req.headers.get(self.auth_header, None) == auth:
return None
req.add_header(self.auth_header, auth)
return self.parent.open(req)
else:
return None
As you can see by inspection, I've just changed two strings from "Basic" to
"NxBasic" (and the lowercase equivalents) from what's in urrlib2.py (in the
abstract basic auth handler superclass of the http basic auth handler class).
Try using this version -- and if it's still not working, at least having it be
your code can help you add print/logging statements, breakpoints, etc, to
better understand what's breaking and how. Best of luck! (Sorry I can't help
further but I don't have any Nexus around to experiment with).
|
How would you adblock using Python?
Question: I'm slowly building a [web
browser](http://github.com/regomodo/qtBrowser/tree/master) in PyQt4 and like
the speed i'm getting out of it. However, I want to combine easylist.txt with
it. I believe adblock uses this to block http requests by the browser.
How would you go about it using python/PyQt4?
[edit1] Ok. I think i've setup Privoxy. I haven't setup any additional filters
and it seems to work. The PyQt4 i've tried to use looks like this
`self.proxyIP = "127.0.0.1"
self.proxyPORT= 8118
proxy = QNetworkProxy()
proxy.setType(QNetworkProxy.HttpProxy)
proxy.setHostName(self.proxyIP)
proxy.setPort(self.proxyPORT)
QNetworkProxy.setApplicationProxy(proxy) `
However, this does absolutely nothing and I cannot make sense of the docs and
can not find any examples.
[edit2] I've just noticed that i'f I change self.proxyIP to my actual local IP
rather than 127.0.0.1 the page doesn't load. So something is happening.
Answer: I know this is an old question, but I thought I'd try giving an answer for
anyone who happens to stumble upon it. You could create a subclass of
QNetworkAccessManager and combine it with
<https://github.com/atereshkin/abpy>. Something kind of like this:
from PyQt4.QtNetwork import QNetworkAccessManager
from abpy import Filter
adblockFilter = Filter(file("easylist.txt"))
class MyNetworkAccessManager(QNetworkAccessManager):
def createRequest(self, op, request, device=None):
url = request.url().toString()
doFilter = adblockFilter.match(url)
if doFilter:
return QNetworkAccessManager.createRequest(self, self.GetOperation, QNetworkRequest(QUrl()))
else:
QNetworkAccessManager.createRequest(self, op, request, device)
myNetworkAccessManager = MyNetworkAccessManager()
After that, set the following on all your QWebView instances, or make a
subclass of QWebView:
QWebView.page().setNetworkAccessManager(myNetworkAccessManager)
Hope this helps!
|
Running JSON through Python's eval()?
Question: Best practices aside, is there a compelling reason **not** to do this?
I'm writing a post-commit hook for use with a Google Code project, which
provides commit data via a JSON object. GC provides an HMAC authentication
token along with the request (outside the JSON data), so by validating that
token I gain high confidence that the JSON data is both benign (as there's
little point in distrusting Google) and valid.
My own (brief) investigations suggest that JSON happens to be completely valid
Python, with the exception of the `"\/"` escape sequence — which GC doesn't
appear to generate.
So, as I'm working with Python 2.4 (i.e. no `json` module), `eval()` is
looking _really_ tempting.
**Edit:** For the record, I am very much **not** asking if this is a good
idea. I'm _quite_ aware that it isn't, and I very much doubt I'll ever use
this technique for any future projects even if I end up using it for this one.
I just wanted to make sure that I know what kind of trouble I'll run into if I
do. :-)
Answer: If you're comfortable with your script working fine for a while, and then
randomly failing on some obscure edge case, I would go with eval.
If it's important that your code be robust, I would take the time to add
simplejson. You don't need the C portion for speedups, so it really shouldn't
be hard to dump a few .py files into a directory somewhere.
As an example of something that might bite you, JSON uses Unicode and
simplejson returns Unicode, whereas eval returns str:
>>> simplejson.loads('{"a":1, "b":2}')
{u'a': 1, u'b': 2}
>>> eval('{"a":1, "b":2}')
{'a': 1, 'b': 2}
Edit: a better example of where eval() behaves differently:
>>> simplejson.loads('{"X": "\uabcd"}')
{u'X': u'\uabcd'}
>>> eval('{"X": "\uabcd"}')
{'X': '\\uabcd'}
>>> simplejson.loads('{"X": "\uabcd"}') == eval('{"X": "\uabcd"}')
False
Edit 2: saw yet another problem today pointed out by SilentGhost: eval doesn't
handle true -> True, false -> False, null -> None correctly.
>>> simplejson.loads('[false, true, null]')
[False, True, None]
>>> eval('[false, true, null]')
Traceback (most recent call last):
File "<interactive input>", line 1, in <module>
File "<string>", line 1, in <module>
NameError: name 'false' is not defined
>>>
|
How do I use TLS with asyncore?
Question: An asyncore-based XMPP client opens a normal TCP connection to an XMPP server.
The server indicates it requires an encrypted connection. The client is now
expected to start a TLS handshake so that subsequent requests can be
encrypted.
[tlslite](http://trevp.net/tlslite/readme.txt) integrates with asyncore, but
the sample code is for a server (?) and I don't understand what it's doing.
I'm on Python 2.5. How can I get the TLS magic working?
* * *
Here's what ended up working for me:
from tlslite.api import *
def handshakeTls(self):
"""
Encrypt the socket using the tlslite module
"""
self.logger.info("activating TLS encrpytion")
self.socket = TLSConnection(self.socket)
self.socket.handshakeClientCert()
Answer: Definitely check out twisted and wokkel. I've been building tons of xmpp bots
and components with it and it's a dream.
|
Python remove all lines which have common value in fields
Question: I have lines of data comprising of 4 fields
aaaa bbb1 cccc dddd
aaaa bbb2 cccc dddd
aaaa bbb3 cccc eeee
aaaa bbb4 cccc ffff
aaaa bbb5 cccc gggg
aaaa bbb6 cccc dddd
Please bear with me.
The first and third field is always the same - but I don't need them, the 4th
field can be the same or different. The thing is, I only want 2nd and 4th
fields from lines which don't share the common field. For example like this
from the above data
bbb3 eeee
bbb4 ffff
bbb5 gggg
Now I don't mean deduplication as that would leave one of the entries in. If
the 4th field shares a value with another line, I don't want any line which
ever had that value.
humblest apologies once again for asking what is probably simple.
Answer: Here you go:
from collections import defaultdict
LINES = """\
aaaa bbb1 cccc dddd
aaaa bbb2 cccc dddd
aaaa bbb3 cccc eeee
aaaa bbb4 cccc ffff
aaaa bbb5 cccc gggg
aaaa bbb6 cccc dddd""".split('\n')
# Count how many lines each unique value of the fourth field appears in.
d_counts = defaultdict(int)
for line in LINES:
a, b, c, d = line.split()
d_counts[d] += 1
# Print only those lines with a unique value for the fourth field.
for line in LINES:
a, b, c, d = line.split()
if d_counts[d] == 1:
print b, d
# Prints
# bbb3 eeee
# bbb4 ffff
# bbb5 gggg
|
How can I parse marked up text for further processing?
Question: **See updated input and output data at Edit-1.**
What I am trying to accomplish is turning
+ 1
+ 1.1
+ 1.1.1
- 1.1.1.1
- 1.1.1.2
+ 1.2
- 1.2.1
- 1.2.2
- 1.3
+ 2
- 3
into a python data structure such as
[{'1': [{'1.1': {'1.1.1': ['1.1.1.1', '1.1.1.2']}, '1.2': ['1.2.1', '1.2.2']}, '1.3'], '2': {}}, ['3',]]
I've looked at many different wiki markup languages, markdown, restructured
text, etc but they are all extremely complicated for me to understand how it
works since they must cover a large amount of tags and syntax (I would only
need the "list" parts of most of these but converted to python instead of html
of course.)
I've also taken a look at tokenizers, lexers and parsers but again they are
much more complicated than I need and that I can understand.
I have no idea where to begin and would appreciate any help possible on this
subject. Thanks
**Edit-1** : Yes the character at the beginning of the line matters, from the
required output from before and now it could be seen that the **`*`** denotes
a root node with children, the **+** has children and the **-** has no
children (root or otherwise) and is just extra information pertaining to that
node. The **`*`** is not important and can be interchanged with **+** (I can
get root status other ways.)
Therefore the new requirement would be using only **`*`** to denote a node
with or without children and **-** cannot have children. I've also changed it
so the key isn't the text after the **`*`** since that will no doubt changer
later to an actual title.
For example
* 1
* 1.1
* 1.2
- Note for 1.2
* 2
* 3
- Note for root
would give
[{'title': '1', 'children': [{'title': '1.1', 'children': []}, {'title': '1.2', 'children': []}]}, {'title': '2', 'children': [], 'notes': ['Note for 1.2', ]}, {'title': '3', 'children': []}, 'Note for root']
Or if you have another idea to represent the outline in python then bring it
forward.
Answer: **Edit** : thanks to the clarification and change in the spec I've edited my
code, still using an explicit `Node` class as an intermediate step for clarity
-- the logic is to turn the list of lines into a list of nodes, then turn that
list of nodes into a tree (by using their indent attribute appropriately),
then print that tree in a readable form (this is just a "debug-help" step, to
check the tree is well constructed, and can of course get commented out in the
final version of the script -- which, just as of course, will take the lines
from a file rather than having them hardcoded for debugging!-), finally build
the desired Python structure and print it. Here's the code, and as we'll see
after that the result is _almost_ as the OP specifies with one exception --
but, the code first:
import sys
class Node(object):
def __init__(self, title, indent):
self.title = title
self.indent = indent
self.children = []
self.notes = []
self.parent = None
def __repr__(self):
return 'Node(%s, %s, %r, %s)' % (
self.indent, self.parent, self.title, self.notes)
def aspython(self):
result = dict(title=self.title, children=topython(self.children))
if self.notes:
result['notes'] = self.notes
return result
def print_tree(node):
print ' ' * node.indent, node.title
for subnode in node.children:
print_tree(subnode)
for note in node.notes:
print ' ' * node.indent, 'Note:', note
def topython(nodelist):
return [node.aspython() for node in nodelist]
def lines_to_tree(lines):
nodes = []
for line in lines:
indent = len(line) - len(line.lstrip())
marker, body = line.strip().split(None, 1)
if marker == '*':
nodes.append(Node(body, indent))
elif marker == '-':
nodes[-1].notes.append(body)
else:
print>>sys.stderr, "Invalid marker %r" % marker
tree = Node('', -1)
curr = tree
for node in nodes:
while node.indent <= curr.indent:
curr = curr.parent
node.parent = curr
curr.children.append(node)
curr = node
return tree
data = """\
* 1
* 1.1
* 1.2
- Note for 1.2
* 2
* 3
- Note for root
""".splitlines()
def main():
tree = lines_to_tree(data)
print_tree(tree)
print
alist = topython(tree.children)
print alist
if __name__ == '__main__':
main()
When run, this emits:
1
1.1
1.2
Note: 1.2
2
3
Note: 3
[{'children': [{'children': [], 'title': '1.1'}, {'notes': ['Note for 1.2'], 'children': [], 'title': '1.2'}], 'title': '1'}, {'children': [], 'title': '2'}, {'notes': ['Note for root'], 'children': [], 'title': '3'}]
Apart from the ordering of keys (which is immaterial and not guaranteed in a
dict, of course), this is _almost_ as requested -- except that here **all**
notes appear as dict entries with a key of `notes` and a value that's a list
of strings (but the notes entry is omitted if the list would be empty, roughly
as done in the example in the question).
In the current version of the question, how to represent the notes is slightly
unclear; one note appears as a stand-alone string, others as entries whose
value is a string (instead of a list of strings as I'm using). It's not clear
what's supposed to imply that the note must appear as a stand-alone string in
one case and as a dict entry in all others, so this scheme I'm using is more
regular; and if a note (if any) is a single string rather than a list, would
that mean it's an error if more than one note appears for a node? In the
latter regard, this scheme I'm using is more general (lets a node have any
number of notes from 0 up, instead of just 0 or 1 as apparently implied in the
question).
Having written so much code (the pre-edit answer was about as long and helped
clarify and change the specs) to provide (I hope) 99% of the desired solution,
I hope this satisfies the original poster, since the last few tweaks to code
and/or specs to make them match each other should be easy for him to do!
|
Finding partial strings in a list of strings - python
Question: I am trying to check if a user is a member of an Active Directory group, and I
have this:
ldap.set_option(ldap.OPT_REFERRALS, 0)
try:
con = ldap.initialize(LDAP_URL)
con.simple_bind_s(userid+"@"+ad_settings.AD_DNS_NAME, password)
ADUser = con.search_ext_s(ad_settings.AD_SEARCH_DN, ldap.SCOPE_SUBTREE, \
"sAMAccountName=%s" % userid, ad_settings.AD_SEARCH_FIELDS)[0][1]
except ldap.LDAPError:
return None
`ADUser` returns a list of strings:
{'givenName': ['xxxxx'],
'mail': ['[email protected]'],
'memberOf': ['CN=group1,OU=Projects,OU=Office,OU=company,DC=domain,DC=com',
'CN=group2,OU=Projects,OU=Office,OU=company,DC=domain,DC=com',
'CN=group3,OU=Projects,OU=Office,OU=company,DC=domain,DC=com',
'CN=group4,OU=Projects,OU=Office,OU=company,DC=domain,DC=com'],
'sAMAccountName': ['myloginid'],
'sn': ['Xxxxxxxx']}
Of course in the real world the group names are verbose and of varied
structure, and users will belong to tens or hundreds of groups.
If I get the list of groups out as `ADUser.get('memberOf')[0]`, what is the
best way to check if any members of a separate list exist in the main list?
For example, the check list would be `['group2', 'group16']` and I want to get
a true/false answer as to whether _any_ of the smaller list exist in the main
list.
Answer: If the format example you give is somewhat reliable, something like:
import re
grps = re.compile(r'CN=(\w+)').findall
def anyof(short_group_list, adu):
all_groups_of_user = set(g for gs in adu.get('memberOf',()) for g in grps(gs))
return sorted(all_groups_of_user.intersection(short_group_list))
where you pass your list such as `['group2', 'group16']` as the first
argument, your `ADUser` dict as the second argument; this returns an
alphabetically sorted list (possibly empty, meaning "none") of the groups,
among those in `short_group_list`, to which the user belongs.
It's probably not much faster to just a bool, but, if you insist, changing the
second statement of the function to:
return any(g for g in short_group_list if g in all_groups_of_user)
might possibly save a certain amount of time in the "true" case (since `any`
short-circuits) though I suspect not in the "false" case (where the whole list
must be traversed anyway). If you care about the performance issue, best is to
benchmark both possibilities on data that's realistic for your use case!
If performance isn't yet good enough (and a bool yes/no is sufficient, as you
say), try reversing the looping logic:
def anyof_v2(short_group_list, adu):
gset = set(short_group_list)
return any(g for gs in adu.get('memberOf',()) for g in grps(gs) if g in gset)
`any`'s short-circuit abilities might prove more useful here (at least in the
"true" case, again -- because, again, there's no way to give a "false" result
without examining ALL the possibilities anyway!-).
|
Can python mechanize handle HTTP auth?
Question: Mechanize (Python) is failing with 401 for me to open http digest URLs. I
googled and tried debugging but no success.
My code looks like this.
import mechanize
project = "test"
baseurl = "http://trac.somewhere.net"
loginurl = "%s/%s/login" % (baseurl, project)
b = mechanize.Browser()
b.add_password(baseurl, "user", "secret", "some Realm")
b.open(loginurl)
Answer: Mechanize claims that the parameters should be uri, username and password as
parameters, but you have four parameters. Four parameters are correct for
urllib2.add_password, but then the first parameter should be the realm, not
the uri.
<http://wwwsearch.sourceforge.net/mechanize/>
I'd try to change that first.
Does trac require digest? if not a next step could be to try using basic auth,
as a test to see if that works, since you can add that with just addHeader:
import base64
from mechanize import Browser
browser = Browser()
browser.addheaders.append(('Authorization', 'Basic %s' % base64.encodestring('%s:%s' % (user, pwd))))
|
IRC Python Bot: Best Way
Question: I want to build a bot that basically does the following:
1. Listens to the room and interacts with users and encourages them to PM the bot.
2. Once a user has PMed the bot engage with the client using various AI techniques.
Should I just use the IRC library or Sockets in python or do I need more of a
bot framework.
What would you do?
Thanks!
Here is the code I'm currently using, however, I haven't gotten it to work.
#!/usr/bin/python
import socket
network = 'holmes.freenet.net'
port = 6667
irc = socket.socket ( socket.AF_INET, socket.SOCK_STREAM )
irc.connect ( ( network, port ) )
irc.send ( 'NICK PyIRC\r\n' )
irc.send ( 'USER PyIRC PyIRC PyIRC :Python IRC\r\n' )
irc.send ( 'JOIN #pyirc\r\n' )
irc.send ( 'PRIVMSG #pyirc :Can you hear me?\r\n' )
irc.send ( 'PART #pyirc\r\n' )
irc.send ( 'QUIT\r\n' )
irc.close()
Answer: Use [Twisted](http://twistedmatrix.com) or
[Asynchat](http://docs.python.org/library/asynchat.html) if you want to have a
sane design. It is possible to just do it with sockets but why bother doing it
from scratch?
|
Tkinter: AttributeError: NoneType object has no attribute get
Question: I have seen a couple of other posts on similar error message but couldn't find
a solution which would fix it in my case.
I dabbled a bit with TkInter and created a very simple UI. The code follows-
from string import *
from Tkinter import *
import tkMessageBox
root=Tk()
vid = IntVar()
def grabText(event):
if entryBox.get().strip()=="":
tkMessageBox.showerror("Error", "Please enter text")
else:
print entryBox.get().strip()
root.title("My Sample")
root.maxsize(width=550, height=200)
root.minsize(width=550, height=200)
root.resizable(width=NO, height=NO)
label=Label(root, text = "Enter text:").grid(row=2,column=0,sticky=W)
entryBox=Entry(root,width=60).grid(row=2, column=1,sticky=W)
grabBtn=Button(root, text="Grab")
grabBtn.grid(row=8, column=1)
grabBtn.bind('<Button-1>', grabText)
root.mainloop()
I get the UI up and running. When I click on the `Grab` button, I get the
following error on the console:
C:\Python25>python.exe myFiles\testBed.py
Exception in Tkinter callback
Traceback (most recent call last):
File "C:\Python25\lib\lib-tk\Tkinter.py", line 1403, in __call__
return self.func(*args)
File "myFiles\testBed.py", line 10, in grabText
if entryBox.get().strip()=="":
AttributeError: 'NoneType' object has no attribute 'get'
The error traces back to `Tkinter.py`.
I'm sure some one might have dealt with this before. Any help is appreciated.
Answer: The `grid` (and `pack`, and `place`) function of the `Entry` object (and of
all other widgets) returns `None`. In python when you do `a().b()`, the result
of the expression is whatever `b()` returns, therefore `Entry(...).grid(...)`
will return `None`.
You should split that onto two lines, like this:
entryBox = Entry(root, width=60)
entryBox.grid(row=2, column=1, sticky=W)
That way, you get your `Entry` reference stored in `entryBox`, and it's laid
out like you expect. This has a bonus side effect of making your layout easier
to understand and maintain, if you collect all of your `grid` and/or `pack`
statements in blocks.
|
Find functions explicitly defined in a module (python)
Question: Ok I know you can use the dir() method to list everything in a module, but is
there any way to see only the functions that are defined in that module? For
example, assume my module looks like this:
from datetime import date, datetime
def test():
return "This is a real method"
Even if i use inspect() to filter out the builtins, I'm still left with
anything that was imported. E.g I'll see:
['date', 'datetime', 'test']
Is there any way to exclude imports? Or another way to find out what's defined
in a module?
Answer: Are you looking for something like this?
import sys, inspect
def is_mod_function(mod, func):
return inspect.isfunction(func) and inspect.getmodule(func) == mod
def list_functions(mod):
return [func.__name__ for func in mod.__dict__.itervalues()
if is_mod_function(mod, func)]
print 'functions in current module:\n', list_functions(sys.modules[__name__])
print 'functions in inspect module:\n', list_functions(inspect)
EDIT: Changed variable names from 'meth' to 'func' to avoid confusion (we're
dealing with functions, not methods, here).
|
Python Lambda Problems
Question: What's going on here? I'm trying to create a list of functions:
def f(a,b):
return a*b
funcs = []
for i in range(0,10):
funcs.append(lambda x:f(i,x))
This isn't doing what I expect. I would expect the list to act like this:
funcs[3](3) = 9
funcs[0](5) = 0
But all the functions in the list seem to be identical, and be setting the
fixed value to be 9:
funcs[3](3) = 27
funcs[3](1) = 9
funcs[2](6) = 54
Any ideas?
Answer: lambdas in python are closures.... the arguments you give it aren't going to
be evaluated until the lambda is evaluated. At that time, i=9 regardless,
because your iteration is finished.
The behavior you're looking for can be achieved with functools.partial
import functools
def f(a,b):
return a*b
funcs = []
for i in range(0,10):
funcs.append(functools.partial(f,i))
|
List of installed fonts OS X / C
Question: I'm trying to programatically get a list of installed fonts in C or Python. I
need to be able to do this on OS X, does anyone know how?
Answer: Python with PyObjC installed (which is the case for Mac OS X 10.5+, so this
code will work without having to install anything):
import Cocoa
manager = Cocoa.NSFontManager.sharedFontManager()
font_families = list(manager.availableFontFamilies())
(based on htw's answer)
|
Is it possible to pass a variable out of a pdb session into the original interactive session?
Question: I am using pdb to examine a script having called `run -d` in an ipython
session. It would be useful to be able to plot some of the variables but I
need them in the main ipython environment in order to do that.
So what I am looking for is some way to make a variable available back in the
main interactive session after I quit pdb. If you set a variable in the
topmost frame it does seem to be there in the ipython session, but this
doesn't work for any frames further down.
Something like `export` in the following:
ipdb> myvar = [1,2,3]
ipdb> p myvar
[1, 2, 3]
ipdb> export myvar
ipdb> q
In [66]: myvar
Out[66]: [1, 2, 3]
Answer: Per ipython's [docs](http://ipython.scipy.org/moin/Documentation), and also a
`run?` command from the ipython prompt,
> after execution, the IPython interactive namespace gets updated with all
> variables defined in the program (except for `__name__` and sys.argv)
By "defined in the program" (a slightly sloppy use of terms), it doesn't mean
"anywhere within any nested functions found there" -- it means "in the
`globals()` of the script/module you're `run`ning. If you're within any kind
of nesting, `globals()['myvar'] = [1,2,3]` should still work fine, just like
your hoped-for `export` would if it existed.
**Edit** : If you're in a different module, you need to set the name in the
globals of your original one -- after an `import sys` if needed,
`sys.modules["originalmodule"].myvar = [1, 2, 3]` will do what you desire.
|
PyQt: event is not triggered, what's wrong with my code?
Question: I'm a Python newbie and I'm trying to write a trivial app with an event
handler that gets activated when an item in a custom QTreeWidget is clicked.
For some reason it doesn't work. Since I'm only at the beginning of learning
it, I can't figure out what I'm doing wrong. Here is the code:
#!/usr/bin/env python
import sys
from PyQt4.QtCore import SIGNAL
from PyQt4.QtGui import QApplication
from PyQt4.QtGui import QMainWindow
from PyQt4.QtGui import QTreeWidget
from PyQt4.QtGui import QTreeWidgetItem
class MyTreeItem(QTreeWidgetItem):
def __init__(self, s, parent = None):
super(MyTreeItem, self).__init__(parent, [s])
class MyTree(QTreeWidget):
def __init__(self, parent = None):
super(MyTree, self).__init__(parent)
self.setMinimumWidth(200)
self.setMinimumHeight(200)
for s in ['foo', 'bar']:
MyTreeItem(s, self)
self.connect(self, SIGNAL('itemClicked(QTreeWidgetItem*, column)'), self.onClick)
def onClick(self, item, column):
print item
class MainWindow(QMainWindow):
def __init__(self, parent = None):
super(MainWindow, self).__init__(parent)
self.tree = MyTree(self)
def main():
app = QApplication(sys.argv)
win = MainWindow()
win.show()
app.exec_()
if __name__ == '__main__':
main()
My initial goal is to make MyTree.onClick() print something when I click a
tree item (and have access to the clicked item in this handler).
Answer: You should have said
self.connect(self, SIGNAL('itemClicked(QTreeWidgetItem*, int)'), self.onClick)
Notice it says **int** rather than _column_ in the first argument to `SIGNAL`.
You also only need to do the `connect` call once for the tree widget, not once
for each node in the tree.
|
How can I find path to given file?
Question: I have a file, for example "something.exe" and I want to find path to this
file
How can I do this in python?
Answer: Perhaps `os.path.abspath()` would do it:
import os
print os.path.abspath("something.exe")
If your `something.exe` is not in the current directory, you can pass any
relative path and `abspath()` will resolve it.
|
Is it possible to go into ipython from code?
Question: For my debugging needs, `pdb` is pretty good. However, it would be _much_
cooler (and helpful) if I could go into `ipython`. Is this thing possible?
Answer: There is an `ipdb` project which embeds iPython into the standard pdb, so you
can just do:
import ipdb; ipdb.set_trace()
It's installable via the usual `easy_install ipdb`.
`ipdb` is pretty short, so instead of easy_installing you can also create a
file `ipdb.py` somewhere on your Python path and paste the following into the
file:
import sys
from IPython.Debugger import Pdb
from IPython.Shell import IPShell
from IPython import ipapi
shell = IPShell(argv=[''])
def set_trace():
ip = ipapi.get()
def_colors = ip.options.colors
Pdb(def_colors).set_trace(sys._getframe().f_back)
|
How can I make setuptools ignore subversion inventory?
Question: When packaging a Python package with a setup.py that uses the setuptools:
from setuptools import setup
...
the source distribution created by:
python setup.py sdist
not only includes, as usual, the files specified in MANIFEST.in, but it also,
gratuitously, includes all of the files that Subversion lists as being version
controlled beneath the package directory. This is vastly annoying. Not only
does it make it difficult to exercise any sort of explicit control over what
files get distributed with my package, but it means that when I build my
package following an "svn export" instead of an "svn checkout", the contents
of my package might be quite different, since without the .svn metadata
setuptools will make different choices about what to include.
My question: how can I turn off this terrible behavior, so that "setuptools"
treats my project the same way whether I'm using Subversion, or version
control it's never heard of, or a bare tree created with "svn export" that
I've created at the end of my project to make sure it builds cleanly somewhere
besides my working directory?
The best I have managed so far is an ugly monkey-patch:
from setuptools.command import sdist
del sdist.finders[:]
But this is Python, not the jungle, so of course I want a better solution that
involves no monkeys at all. How can I tame setuptools, turn off its magic, and
have it behave sensibly by looking at the visible, predictable rules in my
MANIFEST.py instead?
Answer: I know you know much of this, Brandon, but I'll try to give as a complete
answer as I can (although I'm no setuptools gury) for the benefit of others.
The problem here is that setuptools itself involves quite a lot of black
magick, including using an entry point called setuptools.file_finders where
you can add plugins to find files to include. I am, however, at a complete
loss as to how REMOVE plugins from it...
* Quick workaround: svn export your package to a temporary directory and run the setup.py from there. That means you have no svn, so the svn finder finds no files to include. :)
* Longer workaround: Do you really need setuptools? Setuptools have a lot of features, so the answer is likely yes, but mainly those features are depdenencies (so your dependencies get installed by easy_install), namespace packages (foo.bar), and entry points. Namespace packages can actually be created without setuptools as well. But if you use none of these you might actually get away with just using distutils.
* Ugly workaround: The monkeypatch you gave to sdist in your question, which simply makes the plugin not have any finders, and exit quickly.
So as you see, this answer, although as complete as I can make it, is still
embarrassingly incomplete. I can't actually answer your question, though I
think the answer is "You can't".
|
easy, programmable data plotting
Question: I spend most of my time plotting data, but unfortunately I haven't found a
decent solution for my plotting needs. At the moment, the most powerful and
pleasant library I found that performs plotting is matplotlib. The results are
stunning, but I mostly spend my time fighting with the library when trying to
do simple things like having an arrow as I want. SImilar programs like R and
gnuplot produce visually less appealing results, and they are not GUI based.
On the other hand, programs like xmgrace (or better) allow direct manipulation
of the plotted objects and direct feedback, but they fail on two important
points:
1. if my dataset (normally stored in csv files) changes for some reason, I have to reimport it and perform the manipulations again, by hand
2. once I obtain a nice plot setup, the only way I have to recreate the plot is to use a graphical, interactive program. I would like to have the possibility to run a command line utility on my csv files and get the .pdf as a result, with no human intervention.
I still have to find something that provides me both worlds, and it has an
affordable price. Ideally, I would need an interactive GUI program (a la
Origin) to generate matplotlib-based python scripts.
Does anyone have any hints on software that could address my needs on OSX
(preferably) or Linux ?
Answer: You may want to check out [Igor Pro](http://www.wavemetrics.com/). It's quite
old, and quirky but it provides the most advanced plotting system I've found
yet on the Mac. You can modify anything graphically, at a command line or in
script files. The most powerful feature (IMO) is the ability to automatically
generate a script to recreate a figure or to use a figure to create a script
that generates figures _like_ (in style etc.) a particular figure. I use Igor
for all publication figures I produce.
Data is stored in "waves" (translation: vectors) which encapsulate data and
information about the delta between data points (e.g. time step). Figures
reference waves as their data source. When you update a wave (e.g. by re-
importing a CSV file and specifying that the data overwrite specific waves),
all figures that reference that wave are automatically updated.
You can create "layouts" which are page-layouts containing multiple graphs.
These layouts are also automatically updated whenever any of the figures in
the layout are updated (see above). You can add drawing/text/annotations to
either graphs or layout.s
Be warned: Igor Pro's scripting language is something like the bastard child
of VB and Matlab. It makes my eyes bleed. It makes me pray to whatever God
that the pain just end. **But** the entire system is so powerful that it's
worth it.
|
MySQL db problem in Python
Question: For me mysql db has been successfully instaled in my system.I verified through
the following code that it is successfully installed without any errors.
C:\Python26>python
Python 2.6.1 (r261:67517, Dec 4 2008, 16:51:00) [MSC v.1500 32 bit (Intel)] on
win32
Type "help", "copyright", "credits" or "license" for more information.
>>> import MySQLdb
>>>
But when I imported the mysqldb in my script its giving **No module name
MySQLdb.**
Kindly let me know the problem and the solution..
I am using python 2.6 and mysql is 4.0.3 in windows XP.
Thanks in advance...
Answer: 1) Try using your package manager to download **[python-
mysql](http://www.novell.com/products/linuxpackages/server10/i386/python-
mysql.html)** which includes MySQLdb.
2) Ensure `/usr/lib/python2.4/site-packages/` is in your
[PYTHONPATH](http://docs.python.org/using/cmdline.html#envvar-PYTHONPATH),
e.g.:
>>> import sys
>>> from pprint import pprint
>>> pprint(sys.path)
['',
'/usr/lib/python2.4',
'/usr/lib/python2.4/plat-linux2',
'/usr/lib/python2.4/lib-tk',
'/usr/lib/python2.4/site-packages']
3) You seem to be using the correct capitalization in your example, but it
bears mentioning that the module name is case-sensitive, i.e. MySQLdb
(correct) != mysqldb (incorrect).
**Edit** : Looks like nilamo has found the problem. As mentioned in a comment:
you might be running your script with Python 2.6, but MySQLdb is installed in
2.4's site-packages directory.
|
python + Spreadsheet
Question: Can anybody please tell me is there any possible way to connect to spreadsheet
from python? I want to store some data from a form and submit it to google
spreadsheet. Please help on this issue. What steps do I have to follow?
Thanks in advance...
Answer: The easiest way to connect to Google Spreadsheet is by using this [spreadsheet
library](http://burnash.github.com/gspread). These are the steps you need to
follow:
import gspread
# Login with your Google account
gc = gspread.login('[email protected]','password')
# Spreadsheets can be opened by their title in Google Docs
spreadsheet = gc.open("_YOUR_TARGET_SPREADSHEET_")
# Select worksheet by index
worksheet = spreadsheet.get_worksheet(0)
# Update cell with your form value
worksheet.update_cell(1, 2, form_value_1)
|
Create NTFS junction point in Python
Question: Is there a way to create an NTFS junction point in Python? I know I can call
the `junction` utility, but it would be better not to rely on external tools.
Answer: you can use python win32 API modules e.g.
import win32file
win32file.CreateSymbolicLink(srcDir, targetDir, 1)
see
<http://docs.activestate.com/activepython/2.5/pywin32/win32file__CreateSymbolicLink_meth.html>
for more details
if you do not want to rely on that too, you can always use ctypes and directly
call CreateSymbolicLinl win32 API, which is anyway a simple call
here is example call using ctypes
import ctypes
kdll = ctypes.windll.LoadLibrary("kernel32.dll")
kdll.CreateSymbolicLinkA("d:\testdir", "d:\testdir_link", 1)
[MSDN](http://msdn.microsoft.com/en-us/library/aa363866%28VS.85%29.aspx) says
Minimum supported client Windows Vista
|
Hacking JavaScript Array Into JSON With Python
Question: I am fetching a .js file from a remote site that contains data I want to
process as JSON using the simplejson library on my Google App Engine site. The
.js file looks like this:
var txns = [
{ apples: '100', oranges: '20', type: 'SELL'},
{ apples: '200', oranges: '10', type: 'BUY'}]
I have no control over the format of this file. What I did at first just to
hack through it was to chop the `"var txns = "` bit off of the string and then
do a series of `.replace(old, new, [count])` on the string until it looked
like standard JSON:
cleanJSON = malformedJSON.replace("'", '"').replace('apples:', '"apples":').replace('oranges:', '"oranges":').replace('type:', '"type":').replace('{', '{"transaction":{').replace('}', '}}')
So that it now looks like:
[{ "transaction" : { "apples": "100", "oranges": "20", "type": "SELL"} },
{ "transaction" : { "apples": "200", "oranges": "10", "type": "BUY"} }]
How would you tackle this formatting issue? Is there a known way (library,
script) to format a JavaScript array into JSON notation?
Answer: It's not too difficult to write your own little parsor for that using
[PyParsing](http://pyparsing.wikispaces.com/).
import json
from pyparsing import *
data = """var txns = [
{ apples: '100', oranges: '20', type: 'SELL'},
{ apples: '200', oranges: '10', type: 'BUY'}]"""
def js_grammar():
key = Word(alphas).setResultsName("key")
value = QuotedString("'").setResultsName("value")
pair = Group(key + Literal(":").suppress() + value)
object_ = nestedExpr("{", "}", delimitedList(pair, ","))
array = nestedExpr("[", "]", delimitedList(object_, ","))
return array + StringEnd()
JS_GRAMMAR = js_grammar()
def parse(js):
return JS_GRAMMAR.parseString(js[len("var txns = "):])[0]
def to_dict(object_):
return dict((p.key, p.value) for p in object_)
result = [
{"transaction": to_dict(object_)}
for object_ in parse(data)]
print json.dumps(result)
This is going to print
[{"transaction": {"type": "SELL", "apples": "100", "oranges": "20"}},
{"transaction": {"type": "BUY", "apples": "200", "oranges": "10"}}]
You can also add the assignment to the grammar itself. Given there are already
off-the-shelf parsers for it, you should better use those.
|
Python Printing StdOut As It Received
Question: I'm trying to run wrap a simple (windows) command line tool up in a PyQt GUI
app that I am writing. The problem I have is that the command line tool throws
it's progress out to stdout (it's a server reset command so you get
"Attempting to stop" and "Restarting" type output.
What I am trying to do is capture the output so I can display it as part of my
app. I assumed it would be quite simple to do something like the following :
import os
import subprocess as sub
cmd = "COMMAND LINE APP NAME -ARGS"
proc = sub.Popen(cmd, shell=True, stdout=sub.PIPE).stdout
while 1:
line = proc.readline()
if not line:
break
print line
This partially works in that I do get the contents of StdOut but instead of as
the progress messages are sent I get it once the command line application
exits and it seems to flush StdOut in one go.
Is there a simple answer?
Answer: Interactive communication through stdin/stdout is a common problem.
You're in luck though, with PyQt you can use QProcess, as described here:
[http://diotavelli.net/PyQtWiki/Capturing_Output_from_a_Process](http://diotavelli.net/PyQtWiki/Capturing%5FOutput%5Ffrom%5Fa%5FProcess)
|
Is it possible to fetch a https page via an authenticating proxy with urllib2 in Python 2.5?
Question: I'm trying to add authenticating proxy support to an existing script, as it is
the script connects to a https url (with urllib2.Request and urllib2.urlopen),
scrapes the page and performs some actions based on what it has found.
Initially I had hoped this would be as easy as simply adding a
urllib2.ProxyHandler({"http": MY_PROXY}) as an arg to urllib2.build_opener
which in turn is passed to urllib2.install_opener. Unfortunately this doesn't
seem to work when attempting to do a urllib2.Request(ANY_HTTPS_PAGE). Googling
around lends me to believe that the proxy support in urllib2 in python 2.5
does not support https urls. This surprised me to say the least.
There appear to be solutions floating around the web, for example
<http://bugs.python.org/issue1424152> contains a patch for `urllib2` and
`httplib` which purports to solve the issue (when I tried it the issue I began
to get the following error instead: `urllib2.URLError: <urlopen error (1,
'error:140770FC:SSL routines:SSL23_GET_SERVER_HELLO:unknown protocol')>`).
There is a cookbook recipe here <http://code.activestate.com/recipes/456195>
which I am planning to try next. All in all though I'm surprised this isn't
supported "out of the box", which makes me wonder if I'm simply missing out on
an obvious solutions, so in short — has anyone got a simple method for
fetching https pages using an authenticating proxy with urllib2 in Python 2.5?
Ideally this would work:
import urllib2
#perhaps the dictionary below needs a corresponding "https" entry?
#That doesn't seem to work out of the box.
proxy_handler = urllib2.ProxyHandler({"http": "http://user:pass@myproxy:port"})
urllib2.install_opener( urllib2.build_opener( urllib2.HTTPHandler,
urllib2.HTTPSHandler,
proxy_handler ))
request = urllib2.Request(A_HTTPS_URL)
response = urllib2.urlopen( request)
print response.read()
Many Thanks
Answer: You may want to look into [httplib2](http://code.google.com/p/httplib2/). One
of the [examples](http://code.google.com/p/httplib2/wiki/Examples) claims
support for SOCKS proxies if the [socks](http://socksipy.sourceforge.net/)
module is installed.
|
Listing builtin functions and methods (Python)
Question: I have came up with this:
[a for a in dir(__builtins__) if str(type(getattr(__builtins__,a))) == "<type 'builtin_function_or_method'>"]
I know its ugly. Can you show me a better/more pythonic way of doing this?
Answer: There is the [`inspect` module](http://docs.python.org/library/inspect.html):
import inspect
filter(inspect.isbuiltin, (member for name, member in inspect.getmembers(__builtins__)))
Edit: reading the documentation a little more closely, I came up with this
variant that doesn't use `__getattr__`
import inspect
members = (member for name, member in inspect.getmembers(__builtins__))
filter(inspect.isbuiltin, members)
|
Composable Regexp in Python
Question: Often, I would like to build up complex regexps from simpler ones. The only
way I'm currently aware of of doing this is through string operations, e.g.:
Year = r'[12]\d{3}'
Month = r'Jan|Feb|Mar'
Day = r'\d{2}'
HourMins = r'\d{2}:\d{2}'
Date = r'%s %s, %s, %s' % (Month, Day, Year, HourMins)
DateR = re.compile(Date)
Is anybody aware of a different method or a more systematic approach (maybe a
module) in Python to have composable regexps? I'd rather compile each regexp
individually (e.g. for using individual compile options), but then there
doesn't seem to be a way of composing them anymore!?
Answer: You can use Python's formatting syntax for this:
types = {
"year": r'[12]\d{3}',
"month": r'(Jan|Feb|Mar)',
"day": r'\d{2}',
"hourmins": r'\d{2}:\d{2}',
}
import re
Date = r'%(month)s %(day)s, %(year)s, %(hourmins)s' % types
DateR = re.compile(Date)
(Note the added grouping around Jan|Feb|Mar.)
|
How to use a custom site-package using pth-files for Python 2.6?
Question: I'm trying to setup a custom site-package directory (Python 2.6 on Windows
Vista). For example the directory should be '~\lib\python2.6' (
C:\Users\wierob\lib\python2.6). Hence calling 'setup.py install' should copy
packages to C:\Users\wierob\lib\python2.6.
Following the instructions
[here](http://peak.telecommunity.com/DevCenter/EasyInstall#administrator-
installation):
I've created a pth-file in site-packages directory of the Python installation
(C:\Python26\Lib\site-packages). This file contains a single line:
import os, site; site.addsitedir(os.path.expanduser('~/lib/python2.6'))
Additionally I have a pydistutils.cfg my home directory (C:\Users\wierob) that
contains:
[install]
install_lib = ~/lib/python2.6
install_scripts = ~/bin
When I run 'setup.py install' I get the following error message:
C:\Users\wierob\Documents\Python\workspace\rsreader>setup.py install
running install
Checking .pth file support in C:\Users\wierob\lib\python2.6\
C:\Python26\pythonw.exe -E -c pass
TEST FAILED: C:\Users\wierob\lib\python2.6\ does NOT support .pth files
error: bad install directory or PYTHONPATH
You are attempting to install a package to a directory that is not
on PYTHONPATH and which Python does not read ".pth" files from. The
installation directory you specified (via --install-dir, --prefix, or
the distutils default setting) was:
C:\Users\wierob\lib\python2.6\
So it seems that the pth-file does not work. Although, if I enter
site.addsitedir(os.path.expanduser('~/lib/python2.6'))
in an interactive python session the directory is succesfully added to
sys.path.
Any ideas? Thanks.
Answer: The pth-file seems to be ignored if encoded in UTF-8 with BOM.
Saving the pth-file in ANSI or UTF-8 without BOM works.
|
Need assistance with wxPython (newbie)
Question: I need to create what I _think_ should be a simple GUI. I have very little
experience with building GUI's. I'm a visual learner and 'wxPython In Action'
isn't helping me out. I don't learn well by books written by Ph.D.'s. I'm
using Python 2.6. Many of the examples on the Internet don't work in Python
2.6.
I need to create a GUI with 3 columns and some buttons on the bottom.
On the first pass, each of the columns will be just multi-line text input.
I've creating a GUI that did have 3 columns using 3 panels but I couldn't get
the multi-line text input to fill the entire panel. I tried with boxsizer and
flexgridsizer with one panel but again, I couldn't get the multi-line text
input to fill the entire column.
Somewhere, I saw an example of almost exactly what I was looking for but I
either didn't bookmark it or it was in an example and I forget where it is.
This example had 3 columns where each of the columns could be width adjusted
like in a spreadsheet.
I've been at this for quite a few days and I haven't made any progress. What
I'm looking for is something akin to a Sashwindow but with 3 columns.
I've tried multiple panels, boxsizers with flexgridsizers but no luck. I've
gone through all of the wxPython demos and nothing comes closs. Perhaps
because what I looking for is too simple and not worthy of a demo. Some of the
columns in the real program will use selectable lists and grids but first I
need to start with the simplistic possible case.
Can anyone provide a minimalistic program that shows 3 columns with multi-line
text input filling the entire column? I'll figure out how to add the buttons
on the bottom.
Thank you,
Answer: You should take a look at [wxGlade](http://wxglade.sourceforge.net/). It's a
handy little GUI builder you can use to create your UI. After that, you can
also look at the code it generates and go from there.
**Edit:** Okay, here goes:
In wxGlade, create a new frame. Add a horizontal sizer with three slots. Add a
TextCtrl to the first slot. On the Layout page for the text control, check
wxEXPAND and set Proportion to 1; on the Widget page, check wxTE_MULTILINE.
Copy the text control to the clipboard and paste into into the two remaining
slots.
Here's the code that wxGlade generates:
#!/usr/bin/env python
# -*- coding: iso-8859-15 -*-
# generated by wxGlade 0.6.3 on Tue Jul 21 20:00:54 2009
import wx
# begin wxGlade: extracode
# end wxGlade
class MyFrame(wx.Frame):
def __init__(self, *args, **kwds):
# begin wxGlade: MyFrame.__init__
kwds["style"] = wx.DEFAULT_FRAME_STYLE
wx.Frame.__init__(self, *args, **kwds)
self.text_ctrl_1 = wx.TextCtrl(self, -1, "", style=wx.TE_MULTILINE)
self.text_ctrl_1_copy = wx.TextCtrl(self, -1, "", style=wx.TE_MULTILINE)
self.text_ctrl_1_copy_1 = wx.TextCtrl(self, -1, "", style=wx.TE_MULTILINE)
self.__set_properties()
self.__do_layout()
# end wxGlade
def __set_properties(self):
# begin wxGlade: MyFrame.__set_properties
self.SetTitle("frame_1")
# end wxGlade
def __do_layout(self):
# begin wxGlade: MyFrame.__do_layout
sizer_1 = wx.BoxSizer(wx.VERTICAL)
sizer_2 = wx.BoxSizer(wx.HORIZONTAL)
sizer_2.Add(self.text_ctrl_1, 1, wx.EXPAND, 0)
sizer_2.Add(self.text_ctrl_1_copy, 1, wx.EXPAND, 0)
sizer_2.Add(self.text_ctrl_1_copy_1, 1, wx.EXPAND, 0)
sizer_1.Add(sizer_2, 1, wx.EXPAND, 0)
self.SetSizer(sizer_1)
sizer_1.Fit(self)
self.Layout()
# end wxGlade
# end of class MyFrame
if __name__ == "__main__":
app = wx.PySimpleApp(0)
wx.InitAllImageHandlers()
frame_1 = MyFrame(None, -1, "")
app.SetTopWindow(frame_1)
frame_1.Show()
app.MainLoop()
Hope that helps :-)
|
how do I read everything currently in a subprocess.stdout pipe and then return?
Question: I'm using python's subprocess module to interact with a program via the stdin
and stdout pipes. If I call the subprocesses readline() on stdout, it hangs
because it is waiting for a newline.
How can I do a read of all the characters in the stdout pipe of a subprocess
instance? If it matters, I'm running in Linux.
Answer: Someone else appears to have had the same problem, you can see the related
discussion [here](http://stackoverflow.com/questions/375427/non-blocking-read-
on-a-stream-in-python).
If you are running on Linux you can use select to wait for input on the
process' stdout. Alternatively you change the mode of the process' stdout to
non-blocking using
import fcntl, os
fcntl.fcntl(your_process.stdout, fcntl.F_SETFL, os.O_NONBLOCK)
after which you can loop using read() until you encounter a newline character
(if you want to process the output one line at a time).
|
Why can't I import this Zope component in a Python 2.4 virtualenv?
Question: I'm trying to install Plone 3.3rc4 with plone.app.blob and repoze but nothing
I've tried has worked so far. For one attempt I've pip-installed repoze.zope2,
Plone, and plone.app.blob into a virtualenv. I have [this version of
DocumentTemplate](http://svn.zope.org/Zope/trunk/lib/python/DocumentTemplate/?rev=96249)
in the virtualenv's site-packages directory and I'm trying to get it running
in RHEL5.
For some reason when I try to run `paster serve etc/zope2.ini` in this
environment way Python gives the message `ImportError: No module named
DT_Util`? `DT_Util.py` exists in the directory, `__init__.py` is there too,
and the C module it depends on is there. I suspect there's some circular
dependency or failure when importing the C extension. Of course this module
would work in a normal Zope install...
>>> import DocumentTemplate
Traceback (most recent call last):
File "<stdin>", line 1, in ?
File "DocumentTemplate/__init__.py", line 21, in ?
File ".../lib/python2.4/site-packages/DocumentTemplate/DocumentTemplate.py", line 112, in ?
from DT_String import String, File
File ".../lib/python2.4/site-packages/DocumentTemplate/DT_String.py", line 19, in ?
from DocumentTemplate.DT_Util import ParseError, InstanceDict
ImportError: No module named DT_Util
Answer: I must say I doubt DocumentTemplate from Zope will work standalone. You are
welcome to try though. :-)
Note that [DT_Util imports C
extensions](https://github.com/zopefoundation/DocumentTemplate/blob/master/src/DocumentTemplate/DT_Util.py#L32-L34):
from DocumentTemplate.cDocumentTemplate import InstanceDict, TemplateDict
from DocumentTemplate.cDocumentTemplate import render_blocks, safe_callable
from DocumentTemplate.cDocumentTemplate import join_unicode
You'll need to make sure those are compiled. My guess is that importing the
`cDocumentTemplate` module fails and thus the import of `DT_Util` fails.
|
python docstrings
Question: ok so I decided to learn python (perl, c, c++, java, objective-c, ruby and a
bit of erlang and scala under my belt). and I keep on getting the following
error when I try executing this:
Tue Jul 21{stevenhirsch@steven-hirschs-macbook-pro-2}/projects/python:-->./apache_logs.py
File "./apache_logs.py", line 17
print __doc__
^
SyntaxError: invalid syntax
#!/usr/local/bin/python
"""
USAGE:
apache_logs.py
"""
import sys
import os
if __name__ == "__main__":
if not len(sys.argv) > 1:
print __doc__
sys.exit(1)
infile_name = sys.argv[1]
I know it must be something really stupid but I've googled and read the
documentation without finding anything. The docs all seem to state that what
I've coded should work.
Many thanks in advance for your help!!
Answer: What version of Python do you have? In Python 3, [`print` was changed to work
like a function](http://docs.python.org/3.1/whatsnew/3.0.html#print-is-a-
function) rather than a statement, i.e. `print('Hello World')` instead of
`print 'Hello World'`
I can recommend you to keep using Python 2.6 unless you're doing some brand
new production development. Python 3 is still pretty new.
|
Image resizing with django?
Question: I'm new to Django (and Python) and I have been trying to work out a few things
myself, before jumping into using other people's apps. I'm having trouble
understanding where things 'fit' in the Django (or Python's) way of doing
things. What I'm trying to work out is how to resize an image, once it's been
uploaded. I have my model setup nicely and plugged into the admin, and the
image uploads fine to the directory:
from django.db import models
# This is to list all the countries
# For starters though, this will be just United Kingdom (GB)
class Country(models.Model):
name = models.CharField(max_length=120, help_text="Full name of country")
code = models.CharField(max_length=2, help_text="This is the ISO 3166 2-letter country code (see: http://www.theodora.com/country_digraphs.html)")
flag = models.ImageField(upload_to="images/uploaded/country/", max_length=150, help_text="The flag image of the country.", blank=True)
class Meta:
verbose_name_plural = "Countries"
def __unicode__(self):
return self.name
The thing I'm now having trouble with is taking that file and making a new
file into a thumbnail. Like I say, I'd like to know how to do it without using
others' apps (for now). I have got this code from DjangoSnippets:
from PIL import Image
import os.path
import StringIO
def thumbnail(filename, size=(50, 50), output_filename=None):
image = Image.open(filename)
if image.mode not in ('L', 'RGB'):
image = image.convert('RGB')
image = image.resize(size, Image.ANTIALIAS)
# get the thumbnail data in memory.
if not output_filename:
output_filename = get_default_thumbnail_filename(filename)
image.save(output_filename, image.format)
return output_filename
def thumbnail_string(buf, size=(50, 50)):
f = StringIO.StringIO(buf)
image = Image.open(f)
if image.mode not in ('L', 'RGB'):
image = image.convert('RGB')
image = image.resize(size, Image.ANTIALIAS)
o = StringIO.StringIO()
image.save(o, "JPEG")
return o.getvalue()
def get_default_thumbnail_filename(filename):
path, ext = os.path.splitext(filename)
return path + '.thumb.jpg'
...but this has ultimately confused me... As I don't know how this 'fits in'
to my Django app? And really, is it the best solution for simply making a
thumbnail of an image that has been successfully uploaded? Can anyone possibly
show me a good, solid, decent way that a beginner like me can learn to do this
properly? As in, knowing where to put that sort of code (models.py? forms.py?
...) and how it would work in context? ... I just need a bit of help
understanding and working this problem out.
Thank you!
Answer: If it's OK for you, there is a Django application ready, doing exactly what
you want: <https://github.com/sorl/sorl-thumbnail>
|
Randomness in Jython
Question: When using (pseudo) random numbers in Jython, would it be more efficient to
use the Python random module or Java's random class?
Answer: Python's version is much faster in a simple test on my Mac:
jython -m timeit -s "import random" "random.random()"
1000000 loops, best of 3: 0.266 usec per loop
vs
jython -m timeit -s "import java.util.Random; random=java.util.Random()" "random.nextDouble()"
1000000 loops, best of 3: 1.65 usec per loop
Jython version 2.5b3 and Java version 1.5.0_19.
|
Iterative find/replace from a list of tuples in Python
Question: I have a list of tuples, each containing a find/replace value that I would
like to apply to a string. What would be the most efficient way to do so? I
will be applying this iteratively, so performance is my biggest concern.
More concretely, what would the innards of processThis() look like?
x = 'find1, find2, find3'
y = [('find1', 'replace1'), ('find2', 'replace2'), ('find3', 'replace3')]
def processThis(str,lst):
# Do something here
return something
>>> processThis(x,y)
'replace1, replace2, replace3'
Thanks, all!
Answer: You could consider using `re.sub`:
import re
REPLACEMENTS = dict([('find1', 'replace1'),
('find2', 'replace2'),
('find3', 'replace3')])
def replacer(m):
return REPLACEMENTS[m.group(0)]
x = 'find1, find2, find3'
r = re.compile('|'.join(REPLACEMENTS.keys()))
print r.sub(replacer, x)
|
How to exit from Python without traceback?
Question: I would like to know how to I exit from Python without having an traceback
dump on the output.
I still want want to be able to return an error code but I do not want to
display the traceback log.
I want to be able to exit using `exit(number)` without trace but in case of an
Exception (not an exit) I want the trace.
Answer: You are presumably encountering an exception and the program is exiting
because of this (with a traceback). The first thing to do therefore is to
catch that exception, before exiting cleanly (maybe with a message, example
given).
Try something like this in your `main` routine:
import sys, traceback
def main():
try:
do main program stuff here
....
except KeyboardInterrupt:
print "Shutdown requested...exiting"
except Exception:
traceback.print_exc(file=sys.stdout)
sys.exit(0)
if __name__ == "__main__":
main()
|
Good or bad practice in Python: import in the middle of a file
Question: Suppose I have a relatively long module, but need an external module or method
only once.
Is it considered OK to import that method or module in the middle of the
module?
Or should `import`s only be in the first part of the module.
Example:
import string, pythis, pythat
...
...
...
...
def func():
blah
blah
blah
from pysomething import foo
foo()
etc
etc
etc
...
...
...
**Please justify your answer and add links
to[PEP](http://en.wikipedia.org/wiki/Python_%28programming_language%29#Development)s
or relevant sources**
Answer: [PEP 8](http://www.python.org/dev/peps/pep-0008/) authoritatively states:
> Imports are always put at the top of the file, just after any module
> comments and docstrings, and before module globals and constants.
PEP 8 should be the basis of any "in-house" style guide, since it summarizes
what the core Python team has found to be the most effective style, overall
(and with individual dissent of course, as on any other language, but
consensus and the BDFL agree on PEP 8).
|
Hello World from cython wiki not working
Question: I'm trying to follow this tutorial from Cython:
<http://docs.cython.org/docs/tutorial.html#the-basics-of-cython> and I'm
having a problem.
The files are very simple. I have a helloworld.pyx:
print "Hello World"
and a setup.py:
from distutils.core import setup
from distutils.extension import Extension
from Cython.Distutils import build_ext
setup(
cmdclass = {'build_ext': build_ext},
ext_modules = [Extension("helloworld", ["helloworld.pyx"])]
)
and I compile it with the standard command:
python setup.py build_ext --inplace
I got the following error:
running build
running build_ext
building 'helloworld' extension
gcc -pthread -fno-strict-aliasing -DNDEBUG -g -fwrapv -O2 -Wall -Wstrict-prototypes -fPIC -I/usr/include/python2.6 -c helloworld.c -o build/temp.linux-x86_64-2.6/helloworld.o
helloworld.c:4:20: error: Python.h: No such file or directory
helloworld.c:5:26: error: structmember.h: No such file or directory
helloworld.c:34: error: expected specifier-qualifier-list before ‘PyObject’
helloworld.c:121: error: expected specifier-qualifier-list before ‘PyObject’
helloworld.c:139: error: expected ‘)’ before ‘*’ token
helloworld.c:140: error: expected ‘=’, ‘,’, ‘;’, ‘asm’ or ‘__attribute__’ before ‘__pyx_PyInt_AsLongLong’
helloworld.c:141: error: expected ‘=’, ‘,’, ‘;’, ‘asm’ or ‘__attribute__’ before ‘__pyx_PyInt_AsUnsignedLongLong’
helloworld.c:142: error: expected ‘)’ before ‘*’ token
helloworld.c:147: error: expected ‘)’ before ‘*’ token
helloworld.c:148: error: expected ‘)’ before ‘*’ token
helloworld.c:149: error: expected ‘)’ before ‘*’ token
helloworld.c:150: error: expected ‘)’ before ‘*’ token
helloworld.c:151: error: expected ‘)’ before ‘*’ token
helloworld.c:152: error: expected ‘)’ before ‘*’ token
helloworld.c:153: error: expected ‘)’ before ‘*’ token
helloworld.c:154: error: expected ‘)’ before ‘*’ token
helloworld.c:155: error: expected ‘)’ before ‘*’ token
helloworld.c:156: error: expected ‘)’ before ‘*’ token
helloworld.c:157: error: expected ‘)’ before ‘*’ token
helloworld.c:172: error: expected ‘=’, ‘,’, ‘;’, ‘asm’ or ‘__attribute__’ before ‘*’ token
helloworld.c:173: error: expected ‘=’, ‘,’, ‘;’, ‘asm’ or ‘__attribute__’ before ‘*’ token
helloworld.c:174: error: expected ‘=’, ‘,’, ‘;’, ‘asm’ or ‘__attribute__’ before ‘*’ token
helloworld.c:181: error: expected ‘)’ before ‘*’ token
helloworld.c:198: error: expected ‘=’, ‘,’, ‘;’, ‘asm’ or ‘__attribute__’ before ‘*’ token
helloworld.c:200: error: array type has incomplete element type
helloworld.c:221: error: ‘__pyx_kp_1’ undeclared here (not in a function)
helloworld.c:221: warning: excess elements in struct initializer
helloworld.c:221: warning: (near initialization for ‘__pyx_string_tab[0]’)
helloworld.c:221: warning: excess elements in struct initializer
helloworld.c:221: warning: (near initialization for ‘__pyx_string_tab[0]’)
helloworld.c:221: warning: excess elements in struct initializer
helloworld.c:221: warning: (near initialization for ‘__pyx_string_tab[0]’)
helloworld.c:221: warning: excess elements in struct initializer
helloworld.c:221: warning: (near initialization for ‘__pyx_string_tab[0]’)
helloworld.c:221: warning: excess elements in struct initializer
helloworld.c:221: warning: (near initialization for ‘__pyx_string_tab[0]’)
helloworld.c:221: warning: excess elements in struct initializer
helloworld.c:221: warning: (near initialization for ‘__pyx_string_tab[0]’)
helloworld.c:222: warning: excess elements in struct initializer
helloworld.c:222: warning: (near initialization for ‘__pyx_string_tab[1]’)
helloworld.c:222: warning: excess elements in struct initializer
helloworld.c:222: warning: (near initialization for ‘__pyx_string_tab[1]’)
helloworld.c:222: warning: excess elements in struct initializer
helloworld.c:222: warning: (near initialization for ‘__pyx_string_tab[1]’)
helloworld.c:222: warning: excess elements in struct initializer
helloworld.c:222: warning: (near initialization for ‘__pyx_string_tab[1]’)
helloworld.c:222: warning: excess elements in struct initializer
helloworld.c:222: warning: (near initialization for ‘__pyx_string_tab[1]’)
helloworld.c:222: warning: excess elements in struct initializer
helloworld.c:222: warning: (near initialization for ‘__pyx_string_tab[1]’)
helloworld.c:237: error: expected ‘=’, ‘,’, ‘;’, ‘asm’ or ‘__attribute__’ before ‘inithelloworld’
helloworld.c:238: error: expected ‘=’, ‘,’, ‘;’, ‘asm’ or ‘__attribute__’ before ‘inithelloworld’
helloworld.c:305: error: expected ‘=’, ‘,’, ‘;’, ‘asm’ or ‘__attribute__’ before ‘*’ token
helloworld.c:313: error: expected ‘)’ before ‘*’ token
helloworld.c:379:21: error: compile.h: No such file or directory
helloworld.c:380:25: error: frameobject.h: No such file or directory
helloworld.c:381:23: error: traceback.h: No such file or directory
helloworld.c: In function ‘__Pyx_AddTraceback’:
helloworld.c:384: error: ‘PyObject’ undeclared (first use in this function)
helloworld.c:384: error: (Each undeclared identifier is reported only once
helloworld.c:384: error: for each function it appears in.)
helloworld.c:384: error: ‘py_srcfile’ undeclared (first use in this function)
helloworld.c:385: error: ‘py_funcname’ undeclared (first use in this function)
helloworld.c:386: error: ‘py_globals’ undeclared (first use in this function)
helloworld.c:387: error: ‘empty_string’ undeclared (first use in this function)
helloworld.c:388: error: ‘PyCodeObject’ undeclared (first use in this function)
helloworld.c:388: error: ‘py_code’ undeclared (first use in this function)
helloworld.c:389: error: ‘PyFrameObject’ undeclared (first use in this function)
helloworld.c:389: error: ‘py_frame’ undeclared (first use in this function)
helloworld.c:392: warning: implicit declaration of function ‘PyString_FromString’
helloworld.c:399: warning: implicit declaration of function ‘PyString_FromFormat’
helloworld.c:412: warning: implicit declaration of function ‘PyModule_GetDict’
helloworld.c:412: error: ‘__pyx_m’ undeclared (first use in this function)
helloworld.c:415: warning: implicit declaration of function ‘PyString_FromStringAndSize’
helloworld.c:420: warning: implicit declaration of function ‘PyCode_New’
helloworld.c:429: error: ‘__pyx_empty_tuple’ undeclared (first use in this function)
helloworld.c:440: warning: implicit declaration of function ‘PyFrame_New’
helloworld.c:441: warning: implicit declaration of function ‘PyThreadState_GET’
helloworld.c:448: warning: implicit declaration of function ‘PyTraceBack_Here’
helloworld.c:450: warning: implicit declaration of function ‘Py_XDECREF’
helloworld.c: In function ‘__Pyx_InitStrings’:
helloworld.c:458: error: ‘__Pyx_StringTabEntry’ has no member named ‘p’
helloworld.c:460: error: ‘__Pyx_StringTabEntry’ has no member named ‘is_unicode’
helloworld.c:460: error: ‘__Pyx_StringTabEntry’ has no member named ‘is_identifier’
helloworld.c:461: error: ‘__Pyx_StringTabEntry’ has no member named ‘p’
helloworld.c:461: warning: implicit declaration of function ‘PyUnicode_DecodeUTF8’
helloworld.c:461: error: ‘__Pyx_StringTabEntry’ has no member named ‘s’
helloworld.c:461: error: ‘__Pyx_StringTabEntry’ has no member named ‘n’
helloworld.c:461: error: ‘NULL’ undeclared (first use in this function)
helloworld.c:462: error: ‘__Pyx_StringTabEntry’ has no member named ‘intern’
helloworld.c:463: error: ‘__Pyx_StringTabEntry’ has no member named ‘p’
helloworld.c:463: warning: implicit declaration of function ‘PyString_InternFromString’
helloworld.c:463: error: ‘__Pyx_StringTabEntry’ has no member named ‘s’
helloworld.c:465: error: ‘__Pyx_StringTabEntry’ has no member named ‘p’
helloworld.c:465: error: ‘__Pyx_StringTabEntry’ has no member named ‘s’
helloworld.c:465: error: ‘__Pyx_StringTabEntry’ has no member named ‘n’
helloworld.c:476: error: ‘__Pyx_StringTabEntry’ has no member named ‘p’
helloworld.c: At top level:
helloworld.c:485: error: expected ‘)’ before ‘*’ token
helloworld.c:494: error: expected ‘)’ before ‘*’ token
helloworld.c:500: error: expected ‘=’, ‘,’, ‘;’, ‘asm’ or ‘__attribute__’ before ‘__pyx_PyInt_AsLongLong’
helloworld.c:516: error: expected ‘=’, ‘,’, ‘;’, ‘asm’ or ‘__attribute__’ before ‘__pyx_PyInt_AsUnsignedLongLong’
helloworld.c:538: error: expected ‘)’ before ‘*’ token
helloworld.c:553: error: expected ‘)’ before ‘*’ token
helloworld.c:568: error: expected ‘)’ before ‘*’ token
helloworld.c:583: error: expected ‘)’ before ‘*’ token
helloworld.c:598: error: expected ‘)’ before ‘*’ token
helloworld.c:613: error: expected ‘)’ before ‘*’ token
helloworld.c:628: error: expected ‘)’ before ‘*’ token
helloworld.c:643: error: expected ‘)’ before ‘*’ token
helloworld.c:658: error: expected ‘)’ before ‘*’ token
helloworld.c:673: error: expected ‘)’ before ‘*’ token
helloworld.c:688: error: expected ‘)’ before ‘*’ token
error: command 'gcc' failed with exit status 1
I have python and cython installed from Ubuntu 9.04 repositories. I can't
figure why the compiler can't find Python.h.
I tried doing:
cython helloworld.pyx
and then compiling the result manually with gcc:
gcc -shared -pthread -fPIC -fwrapv -O2 -Wall -fno-strict-aliasing -I/usr/include/python2.5 -o helloworld.so helloworld.c
and got the same exact error message.
Any clues?
Answer: Looks like you're missing some package like `python_dev` or the like -- Debian
and derivatives (including Ubuntu) have long preferred to isolate everything
that could possibly be of "developer"'s use from the parts of a package that
are for "everybody"... a philosophical stance I could debate against (and
_have_ debated against, without much practical success, in mahy fora), but one
that, sadly, can't just be ignored:-(
|
python stdout flush and tee
Question: The following code ends with broken pipe when piped into tee, but behave
correctly when not piped :
#!/usr/bin/python
import sys
def testfun():
while 1:
try :
s = sys.stdin.readline()
except(KeyboardInterrupt) :
print('Ctrl-C pressed')
sys.stdout.flush()
return
print s
if __name__ == "__main__":
testfun()
sys.exit()
Expected output :
./bug.py
Ctrl-C pressed
What is observed when piped into tee is either a broken pipe or no output at
all, ie nothing on tee stdout, and nothing in bug.log :
./bug.py | tee bug.log
Traceback (most recent call last):
File "./bug.py", line 14, in <module>
sys.stdout.flush()
IOError: [Errno 32] Broken pipe
What can be the reason for this ?
Answer: Nope, hitting Ctrl-C does NOT terminate both processes. It terminates the tee
process only, the end of the tee process close the pipe between your script
and tee, and hence your script dies with the broken pipe message.
To fix that, tee has an option to pass the Ctrl-C to its previous process in
the pipe: -i
try: man tee
./bug.py
^CCtrl-C pressed
./bug.py | tee log
^CTraceback (most recent call last):
File "./bug.py", line 14, in <module>
testfun()
File "./bug.py", line 9, in testfun
sys.stdout.flush()
IOError: [Errno 32] Broken pipe
./bug.py | tee -i log
^CCtrl-C pressed
|
combine javascript files at deployment in python
Question: I'm trying to reduce the number of scripts included in our website and we use
buildout to handle deployments. Has anybody successfully implemented a method
of combining and compressing scripts with buildout?
Answer: Here's a Python script I made that I use with all my heavy JavaScript
projects. I'm using YUICompressor, but you can change the code to use another
compressor.
import os, os.path, shutil
YUI_COMPRESSOR = 'yuicompressor-2.4.2.jar'
def compress(in_files, out_file, in_type='js', verbose=False,
temp_file='.temp'):
temp = open(temp_file, 'w')
for f in in_files:
fh = open(f)
data = fh.read() + '\n'
fh.close()
temp.write(data)
print ' + %s' % f
temp.close()
options = ['-o "%s"' % out_file,
'--type %s' % in_type]
if verbose:
options.append('-v')
os.system('java -jar "%s" %s "%s"' % (YUI_COMPRESSOR,
' '.join(options),
temp_file))
org_size = os.path.getsize(temp_file)
new_size = os.path.getsize(out_file)
print '=> %s' % out_file
print 'Original: %.2f kB' % (org_size / 1024.0)
print 'Compressed: %.2f kB' % (new_size / 1024.0)
print 'Reduction: %.1f%%' % (float(org_size - new_size) / org_size * 100)
print ''
#os.remove(temp_file)
I use it like this (the below is just a code snippet, and assumes that the
`compress` function exists in the current namespace):
SCRIPTS = [
'app/js/libs/EventSource.js',
'app/js/libs/Hash.js',
'app/js/libs/JSON.js',
'app/js/libs/ServiceClient.js',
'app/js/libs/jquery.hash.js',
'app/js/libs/Application.js',
'app/js/intro.js',
'app/js/jquery-extras.js',
'app/js/settings.js',
'app/js/api.js',
'app/js/game.js',
'app/js/user.js',
'app/js/pages.intro.js',
'app/js/pages.home.js',
'app/js/pages.log-in.js',
'app/js/pages.log-out.js',
'app/js/pages.new-command.js',
'app/js/pages.new-frame.js',
'app/js/pages.not-found.js',
'app/js/pages.register.js',
'app/js/pages.outro.js',
'app/js/outro.js',
]
SCRIPTS_OUT_DEBUG = 'app/js/multifarce.js'
SCRIPTS_OUT = 'app/js/multifarce.min.js'
STYLESHEETS = [
'app/media/style.css',
]
STYLESHEETS_OUT = 'app/media/style.min.css'
def main():
print 'Compressing JavaScript...'
compress(SCRIPTS, SCRIPTS_OUT, 'js', False, SCRIPTS_OUT_DEBUG)
print 'Compressing CSS...'
compress(STYLESHEETS, STYLESHEETS_OUT, 'css')
if __name__ == '__main__':
main()
|
slow sqlite insert using the jdbc drivers in java
Question: I just inserted 1million records into a simple sqlite table with five columns.
It took a whooping 18 hours in java using the jdbc drivers! I did the same
thing in python2.5 and it took less than a minute. The speed for select
queries seem fine. I think this is an issue with the jdbc drivers.
Is there a faster driver for sqlite3 in java?
Speed of inserting large numbers of rows is important for my schema migration
script, and I'd rather not have to use an external script to do the migrations
if I don't have to.
EDIT: fixed with connection.setAutoCommit(false); thanks Mark Rushakoff for
hinting me to the solution :)
Answer: Did you have your queries autocommitted? That could explain why it took so
long. Try wrapping them in a begin / end so that it doesn't have to do a full
commit for every insert.
[This page](http://www.sqlite.org/lang%5Ftransaction.html) explains begin/end
transaction, while the [FAQ](http://www.sqlite.org/faq.html#q19) touches on
inserts/autocommits.
|
What is the most compatible way to install python modules on a Mac?
Question: I'm starting to learn python and loving it. I work on a Mac mainly as well as
Linux. I'm finding that on Linux (Ubuntu 9.04 mostly) when I install a python
module using apt-get it works fine. I can import it with no trouble.
On the Mac, I'm used to using Macports to install all the Unixy stuff.
However, I'm finding that most of the python modules I install with it are not
being seen by python. I've spent some time playing around with PATH settings
and using python_select . Nothing has really worked and at this point I'm not
really understanding, instead I'm just poking around.
I get the impression that Macports isn't universally loved for managing python
modules. I'd like to start fresh using a more "accepted" (if that's the right
word) approach.
**So, I was wondering, what is the method that Mac python developers use to
manage their modules?**
Bonus questions:
Do you use Apple's python, or some other version? Do you compile everything
from source or is there a package manger that works well (Fink?).
Any tips or suggestions here are greatly appreciated. Thanks for your time. :)
Answer: The most popular way to manage python packages (if you're not using your
system package manager) is to use setuptools and easy_install. It is probably
already installed on your system. Use it like this:
easy_install django
easy_install uses the [Python Package Index](http://pypi.python.org) which is
an amazing resource for python developers. Have a look around to see what
packages are available.
A better option is [pip](http://pypi.python.org/pypi/pip), which is gaining
traction, as it attempts to [fix a lot of the
problems](http://www.b-list.org/weblog/2008/dec/14/packaging/) associated with
easy_install. Pip uses the same package repository as easy_install, it just
works better. Really the only time use need to use easy_install is for this
command:
easy_install pip
After that, use:
pip install django
At some point you will probably want to learn a bit about
[virtualenv](http://pypi.python.org/pypi/virtualenv). If you do a lot of
python development on projects with conflicting package requirements,
virtualenv is a godsend. It will allow you to have completely different
versions of various packages, and switch between them easily depending your
needs.
Regarding which python to use, sticking with Apple's python will give you the
least headaches, but If you need a newer version (Leopard is 2.5.1 I believe),
I would go with the
[macports](http://trac.macports.org/browser/trunk/dports/lang/python26/Portfile)
python 2.6.
|
Ubuntu + virtualenv = a mess? virtualenv hates dist-packages, wants site-packages
Question: Can someone please explain to me what is going on with python in ubuntu 9.04?
I'm trying to spin up `virtualenv`, and the `--no-site-packages` flag seems to
do nothing with ubuntu. I installed `virtualenv 1.3.3` with `easy_install`
(which I've upgraded to `setuptools 0.6c9`) and everything seems to be
installed to `/usr/local/lib/python2.6/dist-packages`
I _assume_ that when installing a package using apt-get, it's placed in
`/usr/lib/python2.6/dist-packages/` ?
The issue is, there is a `/usr/local/lib/python2.6/site-packages` as well that
just sits there being empty. It would seem (by looking at the `path` in a
`virtualenv`) that this is the folder virtualenv uses as backup. Thus even
thought I omit `--no-site-packages`, I cant access my local systems packages
from any of my virtualenv's.
So my questions are:
1. How do I get `virtualenv` to point to one of the `dist-packages`?
2. _Which_ dist-packages should I point it to? `/usr/lib/python2.6/dist-packages` or `/usr/local/lib/python2.6/dist-packages/`
3. What is the point of `/usr/lib/python2.6/site-packages`? There's nothing in there!
4. Is it first come first serve on the path? If I have a newer version of package XYZ installed in `/usr/local/lib/python2.6/dist-packages/` and and older one (from ubuntu repos/apt-get) in `/usr/lib/python2.6/dist-packages`, which one gets imported when I `import xyz`? I'm assuming this is based on the path list, yes?
5. Why the hell is this so confusing? Is there something I'm missing here?
6. Where is it defined that `easy_install` should install to `/usr/local/lib/python2.6/dist-packages`?
7. Will this affect `pip` as well?
Thanks to anyone who can clear this up!
Answer: I believe Mike Orr's answer from [the virtual-env mailing
list](http://groups.google.com/group/python-
virtualenv/browse_thread/thread/1412994580f0a2c6/17271fcd6a6e747e?lnk=gst&q=ubuntu+virtualenv#17271fcd6a6e747e)
seems to be the best. Note the OP published this question in both places.
**Original content of mail:**
Years ago Debian created /usr/local/lib/pythonVERSION/site-packages, and
compiled the Python binary to include it in the default search path. Ubuntu
followed Debian's lead as it normally does. The Python developers did not like
this because you'd get interference with a locally-installed
/usr/local/bin/python using the same site-packages directory. Ubuntu finally
decided to abandon site-packages and use dist-packages instead, a name which
they invented so it wouldn't interfere with anything. The loing story is out
there somewhere if you google it, somewhere in the Python bug tracker or
distutils SIG or such.
The system works, at least if you use the Ubuntu virtualenv package. Some
people have had problems using a locally-installed virtualenv on Ubuntu
because the magic sys.path entries weren't being added or something. I'm not
sure about --no-site-packages because I never use that option: I run PIL and
mysqldb from the Ubuntu packages because it can be hard to compile their C
dependencies sometimes. (Need the right header files, Python is ignoring the
header files, etc.)
So Ubuntu Python packages go into /usr/lib/pythonVERSION/dist-packages. Or
that python-support directory for some reason. Locally-installed Python
packages go into /usr/local/lib/pythonVERSION/dist-packages by default.
Whenever I install an Ubuntu 9.04 system I run:
$ sudo apt-get install python-setuptools (6.0c9) $ sudo apt-get install
python-virtualenv (1.3.3) $ sudo easy_install pip $ sudo pip install
virtualenvwrapper
The virtualenvs work fine this way, although I haven't tried --no-site-
packages.
> I'm trying to spin up virtualenv, and the --no-site-packages flag seems to
> do nothing with ubuntu. I installed virtualenv 1.3.3 with easy_install
> (which I've upgraded to setuptools 0.6c9)
These versions are both in Ubuntu 9.04, so you're making it harder on yourself
by installing them locally.
> and everything seems to be installed to /usr/local/lib/python2.6/dist-
> packages
Yes
> I assume that when installing a package using apt-get, it's placed in /
> usr/lib/python2.6/dist-packages/ ?
Yes
> 4. Is it first come first serve on the path? If I have a newer version of
> package XYZ installed in /usr/local/lib/python2.6/dist- packages/ and and
> older one (from ubuntu repos/apt-get) in /usr/lib/ python2.6/dist-packages,
> which one gets imported when I import xyz? I'm assuming this is based on the
> path list, yes?
>
sys.path is scanned in order. The only funny thing is that .pth eggs get put
earlier or later in the path than some people expect. But if you're using pip
for everything it can do (i.e. except to install pip itself, precompiled eggs,
and a snapshot of a local directory that's a copy rather than an egg link),
you won't have many .pth eggs anyway.
> 5. Why the hell is this so confusing? Is there something I'm missing here?
>
It's not well documented. I figured it out by scanning the web.
> 7. Will this affect pip as well?
>
Yes, pip will automatically install to /usr/local/lib/pythonVERSION/site-
packages. Use "pip install -E $VIRTUAL_ENV packagename" to install into a
virtualenv.
|
Django newbie deployment question - ImportError: Could not import settings 'settings'
Question: The app runs fine using django internal server however when I use apache +
mod_python I get the below error
* * *
File "/usr/local/lib/python2.6/dist-packages/django/conf/__init__.py", line 75, in __init__
raise ImportError, "Could not import settings '%s' (Is it on sys.path? Does it have syntax errors?): %s" % (self.SETTINGS_MODULE, e)
ImportError: Could not import settings 'settings' (Is it on sys.path? Does it have syntax errors?): No module named settings
* * *
Here is the needed information
1) Project directory: /root/djangoprojects/mysite
2) directory listing of /root/djangoprojects/mysite
ls -ltr
total 28
-rw-r--r-- 1 root root 546 Aug 1 08:34 manage.py
-rw-r--r-- 1 root root 0 Aug 1 08:34 __init__.py
-rw-r--r-- 1 root root 136 Aug 1 08:35 __init__.pyc
-rw-r--r-- 1 root root 2773 Aug 1 08:39 settings.py
-rw-r--r-- 1 root root 1660 Aug 1 08:53 settings.pyc
drwxr-xr-x 2 root root 4096 Aug 1 09:04 polls
-rw-r--r-- 1 root root 581 Aug 1 10:06 urls.py
-rw-r--r-- 1 root root 314 Aug 1 10:07 urls.pyc
3) App directory : /root/djangoprojects/mysite/polls
4) directory listing of /root/djangoprojects/mysite/polls
ls -ltr
total 20
-rw-r--r-- 1 root root 514 Aug 1 08:53 tests.py
-rw-r--r-- 1 root root 57 Aug 1 08:53 models.py
-rw-r--r-- 1 root root 0 Aug 1 08:53 __init__.py
-rw-r--r-- 1 root root 128 Aug 1 09:02 views.py
-rw-r--r-- 1 root root 375 Aug 1 09:04 views.pyc
-rw-r--r-- 1 root root 132 Aug 1 09:04 __init__.pyc
5) Anywhere in the filesystem running import django in python interpreter
works fine
6) content of httpd.conf
<Location "/mysite">
SetHandler python-program
PythonHandler django.core.handlers.modpython
SetEnv DJANGO_SETTINGS_MODULE settings
PythonOption django.root /mysite
PythonPath "['/root/djangoprojects/', '/root/djangoprojects/mysite','/root/djangoprojects/mysite/polls', '/var/www'] + sys.path"
PythonDebug On
</Location>
7) PYTHONPATH variable is set to
echo $PYTHONPATH
/root/djangoprojects/mysite
8) DJANGO_SETTINGS_MODULE is set to
echo $DJANGO_SETTINGS_MODULE
mysite.settings
9) content of sys.path is
import sys
>>> sys.path
['', '/root/djangoprojects/mysite', '/usr/lib/python2.6', '/usr/lib/python2.6/plat-linux2', '/usr/lib/python2.6/lib-tk', '/usr/lib/python2.6/lib-old', '/usr/lib/python2.6/lib-dynload', '/usr/lib/python2.6/dist-packages', '/usr/local/lib/python2.6/dist-packages']
How do I add settings location to sys.path such that it persistent across
sessions ?
I have read umpteen no of post with people having the same issue it and I have
tried a lot completely beats me as to what I need to do.
Looking for some help.
Thanks in advance Ankur Gupta
Answer: Your apache configuration should look like this:
<Location "/mysite">
SetHandler python-program
PythonHandler django.core.handlers.modpython
SetEnv DJANGO_SETTINGS_MODULE mysite.settings
PythonOption django.root /mysite
PythonPath "['/root/djangoprojects/', '/root/djangoprojects/mysite','/root/djangoprojects/mysite/polls', '/var/www'] + sys.path"
PythonDebug On
</Location>
Note that the sole difference is the "mysite.settings". Don't forget to
restart apache once the config has changed (apache2ctl restart). See
<http://docs.djangoproject.com/en/dev/howto/deployment/modpython/> for more
info.
|
Error drawing text on NSImage in PyObjC
Question: I'm trying to overlay an image with some text using PyObjC, while striving to
answer my question, ["Annotate images using tools built into OS
X"](http://stackoverflow.com/questions/1229171/annotate-images-using-tools-
built-into-os-x). By referencing [CocoaMagic](http://www.rubycocoa.com/cocoa-
magic-for-gruff-graphs/2/), a RubyObjC replacement for
[RMagick](http://rmagick.rubyforge.org/), I've come up with this:
#!/usr/bin/env python
from AppKit import *
source_image = "/Library/Desktop Pictures/Nature/Aurora.jpg"
final_image = "/Library/Desktop Pictures/.loginwindow.jpg"
font_name = "Arial"
font_size = 76
message = "My Message Here"
app = NSApplication.sharedApplication() # remove some warnings
# read in an image
image = NSImage.alloc().initWithContentsOfFile_(source_image)
image.lockFocus()
# prepare some text attributes
text_attributes = NSMutableDictionary.alloc().init()
font = NSFont.fontWithName_size_(font_name, font_size)
text_attributes.setObject_forKey_(font, NSFontAttributeName)
text_attributes.setObject_forKey_(NSColor.blackColor, NSForegroundColorAttributeName)
# output our message
message_string = NSString.stringWithString_(message)
size = message_string.sizeWithAttributes_(text_attributes)
point = NSMakePoint(400, 400)
message_string.drawAtPoint_withAttributes_(point, text_attributes)
# write the file
image.unlockFocus()
bits = NSBitmapImageRep.alloc().initWithData_(image.TIFFRepresentation)
data = bits.representationUsingType_properties_(NSJPGFileType, nil)
data.writeToFile_atomically_(final_image, false)
When I run it, I get this:
Traceback (most recent call last):
File "/Users/clinton/Work/Problems/TellAtAGlance/ObviouslyTouched.py", line 24, in <module>
message_string.drawAtPoint_withAttributes_(point, text_attributes)
ValueError: NSInvalidArgumentException - Class OC_PythonObject: no such selector: set
Looking in the docs for drawAtPoint:withAttributes:, it says, "You should only
invoke this method when an NSView has focus." NSImage is not a subclass of
NSView, but I would hope this would work, and something very similar seems to
work in the Ruby example.
What do I need to change to make this work?
* * *
I rewrote the code, converting it faithfully, line for line, into an
Objective-C Foundation tool. It works, without problems. [I would be happy to
post if here if there is a reason to do so.]
The question then becomes, how does:
[message_string drawAtPoint:point withAttributes:text_attributes];
differ from
message_string.drawAtPoint_withAttributes_(point, text_attributes)
? Is there a way to tell which "OC_PythonObject" is raising the
NSInvalidArgumentException?
Answer: Here are the problems in the above code:
text_attributes.setObject_forKey_(NSColor.blackColor, NSForegroundColorAttributeName)
->
text_attributes.setObject_forKey_(NSColor.blackColor(), NSForegroundColorAttributeName)
bits = NSBitmapImageRep.alloc().initWithData_(image.TIFFRepresentation)
data = bits.representationUsingType_properties_(NSJPGFileType, nil)
->
bits = NSBitmapImageRep.imageRepWithData_(image.TIFFRepresentation())
data = bits.representationUsingType_properties_(NSJPEGFileType, None)
Minor typos indeed.
Note that the middle portion of the code can be replaced with this more-
readable variant:
# prepare some text attributes
text_attributes = {
NSFontAttributeName : NSFont.fontWithName_size_(font_name, font_size),
NSForegroundColorAttributeName : NSColor.blackColor()
}
# output our message
NSString.drawAtPoint_withAttributes_(message, (400, 400), text_attributes)
I learned this by looking at the source code to [NodeBox](http://nodebox.net),
the twelve lines of
[psyphography.py](http://dev.nodebox.net/browser/nodebox/trunk/play/psyphography.py)
and
[cocoa.py](http://dev.nodebox.net/browser/nodebox/trunk/src/nodebox/graphics/cocoa.py),
particularly the save and _getImageData methods.
|
MySQL LOAD DATA LOCAL INFILE example in python?
Question: I am looking for a syntax definition, example, sample code, wiki, etc. for
executing a LOAD DATA LOCAL INFILE command from python.
I believe I can use mysqlimport as well if that is available, so any feedback
(and code snippet) on which is the better route, is welcome. A Google search
is not turning up much in the way of current info
The goal in either case is the same: Automate loading hundreds of files with a
known naming convention & date structure, into a single MySQL table.
David
Answer: Well, using python's MySQLdb, I use this:
connection = MySQLdb.Connect(host='**', user='**', passwd='**', db='**')
cursor = connection.cursor()
query = "LOAD DATA INFILE '/path/to/my/file' INTO TABLE sometable FIELDS TERMINATED BY ';' ENCLOSED BY '\"' ESCAPED BY '\\\\'"
cursor.execute( query )
connection.commit()
replacing the host/user/passwd/db as appropriate for your needs. This is based
on the MySQL docs [here](http://dev.mysql.com/doc/refman/5.1/en/load-
data.html), The exact LOAD DATA INFILE statement would depend on your specific
requirements etc (note the FIELDS TERMINATED BY, ENCLOSED BY, and ESCAPED BY
statements will be specific to the type of file you are trying to read in).
|
How can I use a VB6 COM 'reference' in IronPython?
Question: I'm currently developing what is more or less a script that needs to get some
data from a VB 6 COM dll. This dll is currently used in a MS Word VBA project,
and it exports classes, etc to the VBA code. It is added in the Tools ->
References menu in the VBA editor, and I can see it's classes in the object
browser of VBA.
From my readings, it is possible to use a VB6 COM library in VB.NET (or at
least, it is supposed to be able to.) As it should be possible in VB.NET, and
since .NET runs on the CLR, and since IronPython does to, logically, can't i
access this ancient DLL from IronPython?
I have tried `import clr; clr.AddReferenceToFileAndPath(dllpath)` in
IronPython, but keep getting 'IOError: file does not exist', which is clearly
false.
If anyone can point me to using a VB6 COM object in /any/ .NET language, that
would be greatly appreciated.
Thanks!
PS: No, I can't edit / view source of the COM DLL, it's 3rd party proprietary
stuff. PPS: Any way I can get the GUID / COM 'name; of the dll?
Answer: You need to generate a Runtime COM Wrapper assembly using the `tlbimp` tool,
and add a reference to that; languages which support .net attributes can do
the interoperation explicitly, but even there, autogenerating the wrapper is
far simpler.
Inspecting the wrapper assembly in `ildasm` will show exactly how the
conversion has been performed.
|
How do you USE Fortran 90 module data
Question: Let's say you have a Fortran 90 module containing _lots_ of variables,
functions and subroutines. In your `USE` statement, which convention do you
follow:
1. explicitly declare which variables/functions/subroutines you're using with the `, only :` syntax, such as `USE [module_name], only : variable1, variable2, ...`?
2. Insert a blanket `USE [module_name]`?
On the one hand, the `only` clause makes the code a bit more verbose. However,
it forces you to repeat yourself in the code and if your module contains
_lots_ of variables/functions/subroutines, things begin to look unruly.
Here's an example:
module constants
implicit none
real, parameter :: PI=3.14
real, parameter :: E=2.71828183
integer, parameter :: answer=42
real, parameter :: earthRadiusMeters=6.38e6
end module constants
program test
! Option #1: blanket "use constants"
! use constants
! Option #2: Specify EACH variable you wish to use.
use constants, only : PI,E,answer,earthRadiusMeters
implicit none
write(6,*) "Hello world. Here are some constants:"
write(6,*) PI, &
E, &
answer, &
earthRadiusInMeters
end program test
**Update** Hopefully someone says something like _"Fortran? Just recode it in
C#!"_ so I can down vote you.
* * *
**Update**
I like [Tim Whitcomb's answer](http://stackoverflow.com/questions/1240510/how-
do-you-use-fortran-90-module-data/1240696#1240696), which compares Fortran's
`USE modulename` with Python's `from modulename import *`. A topic which has
been on Stack Overflow before:
* [‘import module’ or ‘from module import’](http://stackoverflow.com/questions/710551/import-module-or-from-module-import)
* [In an answer](http://stackoverflow.com/questions/710551/import-module-or-from-module-import/710603#710603), Mark Roddy mentioned:
> don't use 'from module import *'. For any reasonable large set of code, if
> you 'import *' your will likely be cementing it into the module, unable to
> be removed. This is because it is difficult to determine what items used in
> the code are coming from 'module', making it east to get to the point where
> you think you don't use the import anymore but its extremely difficult to be
> sure.
* [What are good rules of thumb for python imports?](http://stackoverflow.com/questions/193919/what-are-good-rules-of-thumb-for-python-imports)
* [dbr's answer](http://stackoverflow.com/questions/193919/what-are-good-rules-of-thumb-for-python-imports/193979#193979) contains
> don't do from x import * - it makes your code very hard to understand, as
> you cannot easily see where a method came from (from x import *; from y
> import *; my_func() - where is my_func defined?)
So, I'm leaning towards a consensus of explicitly stating all the items I'm
using in a module via
USE modulename, only : var1, var2, ...
And as [Stefano Borini
mentions](http://stackoverflow.com/questions/1240510/how-do-you-use-
fortran-90-module-data/1240583#1240583),
> [if] you have a module so large that you feel compelled to add ONLY, it
> means that your module is too big. Split it.
Answer: I used to just do `use modulename` \- then, as my application grew, I found it
more and more difficult to find the source to functions (without turning to
grep) - some of the other code floating around the office still uses a one-
subroutine-per-file, which has its own set of problems, but it makes it much
easier to use a text editor to move through the code and quickly track down
what you need.
After experiencing this, I've become a convert to using `use`...`only`
whenever possible. I've also started picking up Python, and view it the same
way as `from modulename import *`. There's a lot of great things that modules
give you, but I prefer to keep my global namespace tightly controlled.
|
Generating a graph with multiple (sets of multiple sets of multiple) X-axis data sets
Question: I am looking for a way to generate a graph with multiple sets of data on the
X-axis, each of which is divided into multiple sets of multiple sets. I
basically want to take [this graph](http://gdgraph.com/samples/sample1A.html)
and place similar graphs side by side with it. I am trying to graph the build
a graph of the duration (Y-axis) of the same jobs (0-3) with different
configurations (0-1) on multiple servers (each group with the same 8 jobs).
Hopefully the following diagram will illustrate what I am trying to accomplish
(smaller groupings are separated by pipes, larger groupings by double pipes):
|| 0 1 | 0 1 | 0 1 | 0 1 || 0 1 | 0 1 | 0 1 | 0 1 || 0 1 | 0 1 | 0 1 | 0 1 ||
|| 0 | 1 | 2 | 3 || 0 | 1 | 2 | 3 || 0 | 1 | 2 | 3 ||
|| Server 1 || Server 2 || Server 3 ||
Is this possible with either the GD::Graph Perl module or the matplotlib
Python module? I can't find examples or documentation on this subject for
either.
Answer: Here's some Python code that will produce what you're looking for. (The
example uses 3 configurations rather than 2 to make sure the code was fairly
general.)
import matplotlib.pyplot as plt
import random
nconfigs, njobs, nservers = 3, 4, 4
width = .9/(nconfigs*njobs)
job_colors = [(0,0,1), (0,1,0), (1,0,0), (1,0,1)]
def dim(color, fraction=.5):
return tuple([fraction*channel for channel in color])
plt.figure()
x = 0
for iserver in range(nservers):
for ijob in range(njobs):
for iconfig in range(nconfigs):
color = dim(job_colors[ijob], (iconfig+2.)/(nconfigs+1))
plt.bar(x, 1.+random.random(), width, color=color)
x += width
x += .1
plt.show()
This code is probably fairly transparent. The odd term
`(iconfig+2.)/(nconfigs+1)` is just to dim the colors for the different
configurations, but keep them bright enough so the colors can be
distinguished.
The output looks like:
![alt text](http://i28.tinypic.com/dgrnzk.png)
|
can't edit line in python's command line in Linux
Question: I'm running the Python CLI under Linux:
bla:visualization> python
Python 2.1.1 (#18, Nov 1 2001, 11:15:13)
[GCC egcs-2.91.66 19990314/Linux (egcs-1.1.2 release)] on linux2
Type "copyright", "credits" or "license" for more information.
>>>
For some reason the arrow keys and the delete key don't work:
**delete:**
>>> x^H^H^H
**up arrow:**
>>> x^[[A^[[A
etc...
How can I make these work?
Answer: The basic problem is that your Python installation was likely not compiled
with the `readline` library. You can confirm this by trying to import the
`readline` module:
import readline
You should get an error when you import if `readline` is not present.
If this is the case, there's not much you can do other than recompile Python
with the `readline` library, if you can.
|
Is the default configuration of re incorrect on macbooks? Or have I simply misunderstood something?
Question: Python came pre-installed on my macbook and I have been slowly getting
acquainted with the langauge. However, it seems that my configuration of the
re library is incorrect, or I simply misunderstand something and things are
amiss. Whenever I run a python script with "import re", I recieve the
following error:
Traceback (most recent call last):
File "regex.py", line 2, in <module>
import re
File "/System/Library/Frameworks/Python.framework/Versions/2.5/lib/python2.5/re.py", line 4, in <module>
# re-compatible interface for the sre matching engine
AttributeError: 'module' object has no attribute 'compile'
What gives!
Answer: Pretty mysterious problem, given that line 4 in that file (and many other
lines around that line number) is a comment (indeed the error msg itself shows
that comment line!-) so even with the worst misconfiguration I'd be hard put
to reproduce the problem as given.
Let's try to simplify things and check how they may (or may not) break. Please
open a Terminal, mkdir a new empty directory somewhere and cd into it (so we
know there's no filename conflict wrt modules etc), at the bash prompt `unset
PYTHONPATH` (so we know for sure that isn't interfering), `unset
PYTHONSTARTUP` (ditto); then type the command:
$ python -c'import re; print re.__file__'
It should emit the line:
/System/Library/Frameworks/Python.framework/Versions/2.5/lib/python2.5/re.pyc
does it? If so, then we can keep rooting around to understand what name clash
(or whatever) caused your original problem. If the problem persists under such
"clean" conditions then your system is jinxed and I would reinstal Mac OS X
Leopard if I were in your shoes!
|
Simple python/Regex problem: Removing all new lines from a file
Question: I'm becoming acquainted with python and am creating problems in order to help
myself learn the ins and outs of the language. My next problem comes as
follows:
I have copied and pasted a huge slew of text from the internet, but the copy
and paste added several new lines to break up the huge string. I wish to
programatically remove all of these and return the string into a giant blob of
characters. This is obviously a job for regex (I think), and parsing through
the file and removing all instances of the newline character sounds like it
would work, but it doesn't seem to be going over all that well for me.
Is there an easy way to go about this? It seems rather simple.
Answer: The two main alternatives: read everything in as a single string and remove
newlines:
clean = open('thefile.txt').read().replace('\n', '')
or, read line by line, removing the newline that ends each line, and join it
up again:
clean = ''.join(l[:-1] for l in open('thefile.txt'))
The former alternative is probably faster, but, as always, I strongly
recommend you MEASURE speed (e.g., use `python -mtimeit`) in cases of your
specific interest, rather than just assuming you know how performance will be.
REs are probably slower, but, again: don't guess, MEASURE!
So here are some numbers for a specific text file on my laptop:
$ python -mtimeit -s"import re" "re.sub('\n','',open('AV1611Bible.txt').read())"
10 loops, best of 3: 53.9 msec per loop
$ python -mtimeit "''.join(l[:-1] for l in open('AV1611Bible.txt'))"
10 loops, best of 3: 51.3 msec per loop
$ python -mtimeit "open('AV1611Bible.txt').read().replace('\n', '')"
10 loops, best of 3: 35.1 msec per loop
The file is a version of the KJ Bible, downloaded and unzipped from
[here](http://printkjv.ifbweb.com/AV%5Ftxt.zip) (I do think it's important to
run such measurements on one easily fetched file, so others can easily
reproduce them!).
Of course, a few milliseconds more or less on a file of 4.3 MB, 34,000 lines,
may not matter much to you one way or another; but as the fastest approach is
also the simplest one (far from an unusual occurrence, especially in
Python;-), I think that's a pretty good recommendation.
|
How to include and use .eggs/pkg_resources within a project directory targeting python 2.5.1
Question: I have python .egg files that are stored in a relative location to some .py
code. The problem is, I am targeting python 2.5.1 computers which require my
project be self contained in a folder (hundreds of thousands of OLPC XO 8.2.1
release laptops running Sugar). This means I cannot just ./ez_install to
perform a system-wide setuptools/pkg_resources installation.
Example directory structure:
My Application/
My Application/library1.egg
My Application/libs/library2.egg
My Application/test.py
I am wondering how best to import and use library1 and library2 from within
test.py with no pkg_resources system-wide installation. Is my best option
simply to unzip the .egg files?
Thanks for any tips.
Answer: If you want to be able to use pkg_resources, just copy pkg_resources.py
alongside your application's main script. It's designed to be able to be used
this way as a standalone runtime.
|
get_allowed_auths() in paramiko for authentication types
Question: I am trying to get supported authentication types/methods from a running SSH
server in Python.
I found this method get_allowed_auths() in the ServerInterface class in
Paramiko but I can't understand if it is usable in a simple client-like
snippet of code (I am writing something that accomplish in ONLY this task).
Anyone can suggest me some links with examples, other the distribution
documentation? Maybe any other ideas to do this?
Thanks.
Answer: You can try to authenticate using no authentication, which should always fail,
but the server will then send back the auth types that can continue. There is
an `auth_none()` method provided by `paramiko.Transport` to do this.
import paramiko
import socket
s = socket.socket()
s.connect(('localhost', 22))
t = paramiko.Transport(s)
t.connect()
try:
t.auth_none('')
except paramiko.BadAuthenticationType, err:
print err.allowed_types
|
Multipart form post to google app engine not working
Question: I am trying to post a multi-part form using httplib, url is hosted on google
app engine, on post it says Method not allowed, though the post using urllib2
works. Full working example is attached.
My question is what is the difference between two, why one works but not the
other
1. is there a problem in my mulipart form post code?
2. or the problem is with google app engine?
3. or something else ?
* * *
import httplib
import urllib2, urllib
# multipart form post using httplib fails, saying
# 405, 'Method Not Allowed'
url = "http://mockpublish.appspot.com/publish/api/revision_screen_create"
_, host, selector, _, _ = urllib2.urlparse.urlsplit(url)
print host, selector
h = httplib.HTTP(host)
h.putrequest('POST', selector)
BOUNDARY = '----------THE_FORM_BOUNDARY'
content_type = 'multipart/form-data; boundary=%s' % BOUNDARY
h.putheader('content-type', content_type)
h.putheader('User-Agent', 'Python-urllib/2.5,gzip(gfe)')
content = ""
L = []
L.append('--' + BOUNDARY)
L.append('Content-Disposition: form-data; name="test"')
L.append('')
L.append("xxx")
L.append('--' + BOUNDARY + '--')
L.append('')
content = '\r\n'.join(L)
h.putheader('content-length', str(len(content)))
h.endheaders()
h.send(content)
print h.getreply()
# post using urllib2 works
data = urllib.urlencode({'test':'xxx'})
request = urllib2.Request(url)
f = urllib2.urlopen(request, data)
output = f.read()
print output
Edit: After changing putrequest to request (on Nick Johnson's suggestion), it
works
url = "http://mockpublish.appspot.com/publish/api/revision_screen_create"
_, host, selector, _, _ = urllib2.urlparse.urlsplit(url)
h = httplib.HTTPConnection(host)
BOUNDARY = '----------THE_FORM_BOUNDARY'
content_type = 'multipart/form-data; boundary=%s' % BOUNDARY
content = ""
L = []
L.append('--' + BOUNDARY)
L.append('Content-Disposition: form-data; name="test"')
L.append('')
L.append("xxx")
L.append('--' + BOUNDARY + '--')
L.append('')
content = '\r\n'.join(L)
h.request('POST', selector, content,{'content-type':content_type})
res = h.getresponse()
print res.status, res.reason, res.read()
so now the question remains what is the difference between two approaches and
can first first be made to work?
Answer: [Nick Johnson's](http://stackoverflow.com/users/12030/nick-johnson) answer
Have you tried sending the request with httplib using .request() instead of
.putrequest() etc, supplying the headers as a dict?
it works!
|
Reimport a module in python while interactive
Question: I know it can be done, but I never remember how.
How can you reimport a module in python? The scenario is as follows: I import
a module interactively and tinker with it, but then I face an error. I fix the
error in the .py file and then I want to reimport the fixed module without
quitting python. How can I do it ?
Answer: This should work:
reload(my.module)
From the [Python docs](http://docs.python.org/library/functions.html#reload)
> Reload a previously imported module. The argument must be a module object,
> so it must have been successfully imported before. This is useful if you
> have edited the module source file using an external editor and want to try
> out the new version without leaving the Python interpreter.
|
Can you get more information about the online file?
Question: I have a online file: <http://dl_dir.qq.com/qqfile/tm/TM2009Beta_chs.exe>
,please donot download it, i want to determine the software version whether is
changed, so i want more information about it. for example, using python,i can
get this:
import urllib2,urllib
req = urllib2.Request('http://dl_dir.qq.com/qqfile/tm/TM2009Beta_chs.exe')
response = urllib2.urlopen(req)
print response.info()
print response.geturl()
Content-Length: 16868680
Server: qqdlsrv(1.84 for linux)
Connection: close
Content-Disposition: attachment; filename=TM2009Beta_chs.exe
Accept-Ranges: bytes
Content-Type: application/octet-stream
http://dl_dir.qq.com/qqfile/tm/TM2009Beta_chs.exe
Can you get more imformation to let me determine the software version is
changed?
Answer: 1. Download the first thousand bytes or so of the file using the range header.
2. Use [pefile](http://code.google.com/p/pefile/) to parse the PE header and extract version information.
3. With the data, [extract useful information](http://books.google.com/books?id=Pas0YoxygnkC&pg=PA40&lpg=PA40&dq=pe+header+layout&source=bl&ots=QmWK2D3niq&sig=FEWTlTHlaukrlQ-mHybNqW81BjY&hl=en&ei=LuGASpnPH42YsgOj5JH9CA&sa=X&oi=book%5Fresult&ct=result&resnum=3#v=onepage&q=pe%20header%20layout&f=false) such as the time date stamp and other goodies that let you find changes in files without reading the whole thing.
|