text
stringlengths
454
608k
url
stringlengths
17
896
dump
stringclasses
91 values
source
stringclasses
1 value
word_count
int64
101
114k
flesch_reading_ease
float64
50
104
Use keyring for all config based passwords/users Instead of having config files with passwords/users in them (which is bad) there should be some integration with the standard keyring providers in various operating systems, this python project @ http:// Blueprint information - Status: - Not started - Approver: - Mark McLoughlin - Priority: - Low - Drafter: - Yahoo Openstackers! - Direction: - Approved - Assignee: - None - Definition: - Approved - Series goal: - None - Implementation: Unknown - Milestone target: - None - Started by - - Completed by - Related branches Related bugs Sprints Whiteboard Certainly seems like a reasonable idea, but we need a much more detailed plan and proof-of-concept code to make progress here. -- Grizzly design summit session: https:/ Conclusion was to discuss on the mailing list and prove the idea in Nova first. Once proven and other projects are keen to adopt it, we can move it into Oslo. -- I've got a working prototype that I'm hoping to check in shortly. Here's the basic approach. In openstack/ 1. Define a 'secure_source' option (No default). If not specified, then none of the secure actions will be taken. If specified, we assume it's a keyring class that supports 'get_password'. IOW, leave this out of your nova.conf and things will behave just as they always have. 2. Add a new parameter to Opts __init__ method, 'secure' (default to False). 'secure' will also imply 'secret' to make sure it doesn't get logged. 3. In class ConfigOpts, method _do_get, before we check the config parser, if an option is defined as secure *and* we have a defined secure source, then use the secure source's get_password method to get the value. If there's no value, then it's an error. If an option is not secure or there's no secure source, then just proceed with using the config file. In the config file: 1. Specify a secure_source E.g.... secure_source = keyring. 2. Use nested variables for things like db passwords in sql_connection. E.g... sql_connection = mysql:/ In the user module: 1. Define the nested password config options as secure. E.g. cfg. 2. (Optional) Provide your own keyring extension. Before running: 1. Install/configure keyring and populate the key file with the keys and passwords. Following these steps, I was able to have the Nova process talk to MySQL without having the password embedded in nova.conf's sql_connection setting. Thoughts? I hope to have the prototype code checked in for your comments early next week. -- I've suggested in https:/ from openstack.common import cfg from openstack.common import keyring opts = [ cfg. default= secret=True), ] CONF. KEYRING = keyring. class Connection(object): def __init__(self): params = { } -- I still have a lingering concern about this. It seems like people are unlikely to use any of the stock python-keyring backends with OpenStack. Which means this is purely about allowing people to add proprietary backends. That's fine in itself, but it just means the feature will see limited real-world testing.
https://blueprints.launchpad.net/oslo-incubator/+spec/pw-keyrings
CC-MAIN-2020-05
refinedweb
489
66.94
4.2.2. Native contacts analysis — MDAnalysis.analysis.contacts¶ This module contains classes to analyze native contacts Q over a trajectory. Native contacts of a conformation are contacts that exist in a reference structure and in the conformation. Contacts in the reference structure are always defined as being closer then a distance radius. The fraction of native contacts for a conformation can be calculated in different ways. This module supports 3 different metrics listed below, as well as custom metrics. - Hard Cut: To count as a contact the atoms i and j have to be at least as close as in the reference structure. - Soft Cut: The atom pair i and j is assigned based on a soft potential that is 1 if the distance is 0, 1/2 if the distance is the same as in the reference and 0 for large distances. For the exact definition of the potential and parameters have a look at function soft_cut_q(). - Radius Cut: To count as a contact the atoms i and j cannot be further apart than some distance radius. The “fraction of native contacts” Q(t) is a number between 0 and 1 and calculated as the total number of native contacts for a given time frame divided by the total number of contacts in the reference structure. 4.2.2.1. Examples for contact analysis¶ 4.2.2.1.1. One-dimensional contact analysis¶ As an example we analyze the opening (“unzipping”) of salt bridges when the AdK enzyme opens up; this is one of the example trajectories in MDAnalysis. import numpy as np import matplotlib.pyplot as plt import MDAnalysis as mda from MDAnalysis.analysis import contacts from MDAnalysis.tests.datafiles import PSF,DCD # example trajectory (transition of AdK from closed to open) u = mda.Universe(PSF,DCD) #) # set up analysis of native contacts ("salt bridges"); salt bridges have a # distance <6 A ca1 = contacts.Contacts(u, select=(sel_acidic, sel_basic), refgroup=(acidic, basic), radius=6.0) # iterate through trajectory and perform analysis of "native contacts" Q ca1.run() # print number of averave contacts average_contacts = np.mean(ca1.timeseries[:, 1]) print('average contacts = {}'.format(average_contacts)) # plot time series q(t) fig, ax = plt.subplots() ax.plot(ca1.timeseries[:, 0], ca1.timeseries[:, 1]) ax.set(xlabel='frame', ylabel='fraction of native contacts', title='Native Contacts, average = {:.2f}'.format(average_contacts)) fig.show() The first graph shows that when AdK opens, about 20% of the salt bridges that existed in the closed state disappear when the enzyme opens. They open in a step-wise fashion (made more clear by the movie AdK_zipper_cartoon.avi). Notes Suggested cutoff distances for different simulations - For all-atom simulations, cutoff = 4.5 Å - For coarse-grained simulations, cutoff = 6.0 Å 4.2.2.1.2. Two-dimensional contact analysis (q1-q2)¶ Analyze a single DIMS transition of AdK between its closed and open conformation and plot the trajectory projected on q1-q2 [Franklin2007] import MDAnalysis as mda from MDAnalysis.analysis import contacts from MDAnalysisTests.datafiles import PSF, DCD u = mda.Universe(PSF, DCD) q1q2 = contacts.q1q2(u, 'name CA', radius=8) q1q2.run() f, ax = plt.subplots(1, 2, figsize=plt.figaspect(0.5)) ax[0].plot(q1q]. 4.2.2.1.3. Writing your own contact analysis¶ The Contacts class has been designed to be extensible for your own analysis. As an example we will analyze when the acidic and basic groups of AdK are in contact which each other; this means that at least one of the contacts formed in the reference is closer than 2.5 Å. For this we define a new function to determine if any contact is closer than 2.5 Å; this function must implement the API prescribed by Contacts: def is_any_closer(r, r0, dist=2.5): return np.any(r < dist) The first two parameters r and r0 are provided by Contacts when it calls is_any_closer() while the others can be passed as keyword args using the kwargs parameter in Contacts. Next we are creating an instance of the Contacts class and use the is_any_closer() function as an argument to method and run the analysis: #) nc = contacts.Contacts(u, select=(sel_acidic, sel_basic), method=is_any_closer, refgroup=(acidic, basic), kwargs={'dist': 2.5}) nc.run() bound = nc.timeseries[:, 1] frames = nc.timeseries[:, 0] f, ax = plt.subplots() ax.plot(frames, bound, '.') ax.set(xlabel='frame', ylabel='is Bound', ylim=(-0.1, 1.1)) f.show() 4.2.2.2. Functions¶ MDAnalysis.analysis.contacts. hard_cut_q(r, cutoff)[source]¶ Calculate fraction of native contacts Q for a hard cut off. The cutoff can either be a float or a ndarrayof the same shape as r. MDAnalysis.analysis.contacts. soft_cut_q(r, r0, beta=5.0, lambda_constant=1.8)[source]¶ Calculate fraction of native contacts Q for a soft cut off The native contact function is defined as [Best2013]\[Q(r, r_0) = \frac{1}{1 + e^{\beta (r - \lambda r_0)}}\] Reasonable values for different simulation types are - All Atom: lambda_constant = 1.8 (unitless) - Coarse Grained: lambda_constant = 1.5 (unitless) References MDAnalysis.analysis.contacts. radius_cut_q(r, r0, radius)[source]¶ calculate native contacts Q based on the single distance radius. References MDAnalysis.analysis.contacts. contact_matrix(d, radius, out=None)[source]¶ calculate contacts from distance matrix MDAnalysis.analysis.contacts. q1q2(u, select='all', radius=4.5)[source]¶ Perform a q1-q2 analysis. Compares native contacts between the starting structure and final structure of a trajectory [Franklin2007]. Changed in version 1.0.0: Changed selection keyword to select Support for setting start, stop, and stephas been removed. These should now be directly passed to Contacts.run(). 4.2.2.3. Classes¶ - class MDAnalysis.analysis.contacts. Contacts(u, select, refgroup, method='hard_cut', radius=4.5, pbc=True, kwargs=None, **basekwargs)[source]¶ Calculate contacts based observables. The standard methods used in this class calculate the fraction of native contacts Q from a trajectory. Contact API By defining your own method it is possible to calculate other observables that only depend on the distances and a possible reference distance. The Contact API prescribes that this method must be a function with call signature func(r, r0, **kwargs)and must be provided in the keyword argument method. .. versionchanged:: 1.0.0 save()method has been removed. Use np.savetxt()on Contacts.timeseriesinstead. .. versionchanged:: 1.0.0 added pbcattribute to calculate distances using PBC. Notes Changed in version 1.0.0: Changed selection keyword to select
https://docs.mdanalysis.org/2.0.0-dev0/documentation_pages/analysis/contacts.html
CC-MAIN-2020-50
refinedweb
1,063
50.84
.NET Back to Basics: The String Class WEBINAR: On-Demand Full Text Search: The Key to Better Natural Language Queries for NoSQL in Node.js Last month, we went back to basics with the Int class. This month, we look at the textual equivalent, the string class. The String & Int data types, single-handed, are the two most used data in the .NET platform. Between them, they handle about 90% of all the data we use. Strings are nothing complex; they are just simple sequences of characters that make up words and sentences. The string class exists to allow us to chop up these sentences, replace parts of them, search them, and a whole bunch of other things that make handling strings of text easier for us as developers. So, what can a string do? The string class has a phenomenal amount of functionality in it, all of which is grouped into about three categories as follows: - String tests - Searching - String modification and building We'll start with "String Tests." When we talk about string tests, we are in fact talking about checking to see if a string is present, or if that string has a certain sub string contained within it. Many developers will instantly recognize that, in this case, an appropriate test would be: if (myname == "shawty") { } String, like most classes that are also basic data types, implements the equality operator, allowing you to perform simple straight-forward tests like this inline in your application code. There is, however, a number of useful specialist tests too. Consider this fragment of code: if (myname == "" && myname == null) { } In language terms, we are saying, IF the variable myname has no contents, that is it is empty and if it is null, that is devoid of any value, then consider our decision logic to be true. The string class makes this much easier with a static method called 'IsNullOrEmpty' if (String.IsNullOrEmpty(myname)) { } To me, the string method version reads much better, and makes better sense from a syntax point of view because the name of the expression tells you the exact test you're performing. Other tests are usually performed as an extension to the string itself. For example 'contains': string myname = "shawty"; if (myname.Contains("shaw")) { } Will pass, because 'shawty' does contain 'shaw'. string myname = "shawty"; if (myname.Contains("peter")) { } will, however, fail. Rather than looking in the string, if you're looking for prefixes or suffixes, 'StartsWith' and 'EndsWith' have your back covered. string myname = "Mr Peter Shaw"; if (myname.StartsWith("Mr") && myname.EndsWith("Shaw")) { } will become true if the first two characters are equal to "Mr" AND the last four are equal to "Shaw" but won't care about any of the other contents in the string. It's also a point to note that the test IS case sensitive, so "Mr" will never match "mr." I soon will show you a way to deal with this, however. All of these tests can also be negated, so to test if a string does NOT start with "Mr" it's as simple as prefixing the method call with an exclamation mark. string myname = "Mr Peter Shaw"; if (!myname.StartsWith("Mr")) { } The string class also contains a 'Compare' method. Compare, in this case, has 11 different versions that operate in subtly different ways, the most simple of which is using System; namespace StringClass { class Program { static void Main(string[] args) { string stringOne = "Peter"; string stringTwo = "shaw"; string stringThree = "Peter"; int compareResult = String.Compare(stringOne, stringTwo); Console.WriteLine("Result was : {0}", compareResult); compareResult = String.Compare(stringOne, stringThree); Console.WriteLine("Result was : {0}", compareResult); compareResult = String.Compare(stringTwo, stringOne); Console.WriteLine("Result was : {0}", compareResult); } } }. Something along the same lines as the following: Figure 1: The output from String.Compare Personally, I find it a bit weird doing comparisons this way, but it does have one advantage that using the standard checks don't have. This is best explained if you look at this MSDN page. Figure 2: The MSDN Docs for compare You'll see I've highlighted the Boolean options available on some of the overrides. If you look at the descriptions, you'll see the entries I've highlighted all state that they can be told to ignore the case of the strings being compared. The other parameters, of which there are many (and for which I would encourage you to read the MSDN docs fully), allow you to use things like culture-specific information (so you can accurately compare things like currency and date formats) or to start the conversion from a given offset in the source or target strings. It's very easy to put together your own extra methods which emulate 'StartsWith', 'EndsWith', and 'Contains' but which are case insensitive. The key to getting it right is just a little experimentation. Searching within a string generally comes in two flavours. The first is via the 'IndexOf', 'LastIndexOf', and the 'Substring' methods. Some of you will undoubtedly be sitting there thinking what? How does IndexOf and Substring constitute searching? Well, technically, many of you might be correct in thinking that IndexOf is really a "String Test" and Substring is well modification to some. When you use them together, however, you use one to find the start of your search, and the second to extract it. This, however, is purely an academic point, which to me makes sense, and because I never use one without the other, is my preferred way of working. So how do you search…? Quite easily. using System; namespace StringClass { class Program { static void Main(string[] args) { string stringOne = "Peter 'Shawty' Shaw"; string stringToSearchFor = "Shawty"; int searchPosition = stringOne.IndexOf("Shawty", 0); string foundString = stringOne.Substring(searchPosition, stringToSearchFor.Length); Console.WriteLine("Found {0} at position {1}", foundString, searchPosition); } } } StringOne is the string to search in, and we use IndexOf on that string to find the index of the string to search for. We then use a substring starting at the position the string to search for was located at and for the length of the string we wish to search for. The second method we have of searching through strings is to use regular expressions. The regex class, however, is a complete class in itself, so we'll devote an entire post to that, at a later date. That brings us to the final group of functionality, "String Modification." Because the string class supports arithmetic operators, you easily can join two string using a + as follows: string name = "Peter " + "Shaw"; Which will result in 'Peter Shaw' in one string. You also can use the concat method. string name = String.Concat("Peter ", "Shaw"); Concat doesn't offer anything over the plus operator, and it only goes up to four parameters, whereas + is infinite. Where it does throw a lifeline, however, is with string lists. using System; using System.Collections.Generic; namespace StringClass { class Program { static void Main(string[] args) { List<string> myStrings = new List<string> { "Peter ", "'Shawty' ", "Shaw ", "With DOT-NET ", "Nuts & Bolts" }; string text = String.Concat(myStrings); Console.WriteLine(text); } } } Which should give you: Figure 3: Output from string list program When used this way, the list length can be practically endless, allowing you to get a large list of strings and combine them to one string. String.Format comes in handy when you want to use a template string and insert sub strings in place holders. You've already seen this done elsewhere in this post. using System; namespace StringClass { class Program { static void Main(string[] args) { string stringOne = "Peter"; string stringTwo = "'Shawty'"; string stringThree = "Shaw"; string stringFour = "DOT-NET"; string stringFive = "Nuts & Bolts"; string text = String.Format("My name is {0} {1} {2} , welcome to {3} {4}", stringOne, stringTwo, stringThree, stringFour, stringFive); Console.WriteLine(text); } } } Many other system routines accept strings in this "format token manner;" for example: the Console.Write and WriteLine methods, which means you don't have to use String.Format directly. String.Join, like concat, can take a list of strings, but this time it joins them using a known separator: using System; using System.Collections.Generic; namespace StringClass { class Program { static void Main(string[] args) { List<string> myStrings = new List<string> { "Peter ", "'Shawty' ", "Shaw ", "With DOT-NET ", "Nuts & Bolts" }; string text = String.Join("|", myStrings); Console.WriteLine(text); } } } As you can see, the first parameter to this is the Pipe character. When you run the program, you should see that all the items in the list have been concatenated with a pipe character between them. Figure 4: All string items have been concatenated, with a pipe character added There's much more the string class can do, but for now I've hit the word limit on this months post. We may have to revisit this again soon. Have Fun! Shawty GhostPosted by Wyze Wildfire on 02/15/2016 07:50am "considered different" vs. "relative position in the sort order" Thank you for the breakdown of information. I noticed that in the middle of your article you wrote the following: ." This one little piece of information bugged me as lacking definition in that it is not explained what defines the decision as to which one is "considered different" as you yourself put it. So I ran some tests, and I believe a more accurate statement would be which one is returned first alphabetically or as Microsoft has stated in their Description of this specific use of this specific method on the page that you have referenced and linked to, "Compares two specified String objects and returns an integer that indicates their relative position in the sort order." So a more clear statement would be: "If the first string would be returned first in an alphabetical sort order then the returning integer = -1 whereas if the first string would be returned second in an alphabetical sort order then the resulting integer returned = 1 For anybody else curious about the defining logic as to how which string is "considered different". Thanks again for the article.Reply
https://www.codeguru.com/columns/dotnet/.net-back-to-basics-the-string-class.html
CC-MAIN-2018-13
refinedweb
1,665
61.36
The problem “Find elements which are present in first array and not in second” states that you are given two arrays. Arrays consist of all the integers. You have to find out the numbers which will not be present in the second array but present in the first array. Example a [] = {2,4,3,1,5,6} b [] = {2,1,5,6} 4 3 a [] ={4,2,6,8,9,5} b [] ={9,3,2,6,8} 4 Algorithm - Declare a HashSet. - Insert all the elements of array b[] into HashSet. - While i < l1 (length of an array a[]). - If HashSet doesn’t contain array a[i], then print a[i]. Explanation We have given two integer arrays and a problem statement that asks to find out the number which is present in the first array and not in the second array. We are going to use Hashing in this problem. Hashing helps us to find out the solution in an efficient way. We are going to put the array b[] numbers in a HashSet and after inserting all the number of array b[]. We are going to traverse array a[] and taking each element at a time and check if HashSet doesn’t contain that element. If it does not have that element, we are going to print that particular element of array a[i] and check for another number. Let us consider an example and understand this: First array is a[]=a [] ={2,6,8,9,5,4}, b [] ={9,5,2,6,8} We have to insert all the elements of array b[] into HashSet, so in HashSet, we have the following values: HashSet:{9,5,2,6,8} // basically all the values of b[]. We will traverse the array a[] and take each of its elements and check the condition. i=0, a[i]=2 2 is in the HashSet, so it will not print. i=1, a[i]=6 6 is in the HashSet, again it will not be printed. i=2, a[i]=8 8 is in the HashSet, it will not be printed. i=3, a[i]=9 9 is in the HashSet, so it will not print. i=4, a[i]=5 5 is in the HashSet, again it will not be printed. i=5, a[i]=4 4 is not in the HashSet, so this time it will be printed means it is the number which is present in an array a[] but not in array b[] because basically HashSet is the clone of array b[] and our output will become ‘4’. C++ code to Find elements which are present in first array and not in second #include<unordered_set> #include<iostream> using namespace std; void getMissingElement(int A[], int B[], int l1, int l2) { unordered_set <int> myset; for (int i = 0; i < l2; i++) myset.insert(B[i]); for (int j = 0; j < l1; j++) if (myset.find(A[j]) == myset.end()) cout << A[j] << " "; } int main() { int a[] = { 9, 2, 3, 1, 4, 5 }; int b[] = { 2, 4, 1, 9 }; int l1 = sizeof(a) / sizeof(a[0]); int l2 = sizeof(b) / sizeof(b[0]); getMissingElement(a, b, l1, l2); return 0; } 3 5 Java code to Find elements which are present in first array and not in second import java.util.HashSet; import java.util.Set; class missingElement { public static void getMissingElement(int A[], int B[]) { int l1 = A.length; int l2 = B.length; HashSet<Integer> set = new HashSet<>(); for (int i = 0; i < l2; i++) set.add(B[i]); for (int i = 0; i < l1; i++) if (!set.contains(A[i])) System.out.print(A[i]+" "); } public static void main(String []args) { int a[] = { 9, 2, 3, 1, 4, 5 }; int b[] = { 2, 4, 1, 9 }; getMissingElement(a, b); } } 3 5 Complexity Analysis Time Complexity O(N) where “N” is the number of elements in the array1. Because using HashSet for insertion and searching allows us to perform these operations in O(1). Thus the time complexity is linear. Space Complexity O(N) where “N” is the number of elements in the array1. Since we are storing the elements of the second array. Thus the space required is the same as that of the size of the second array.
https://www.tutorialcup.com/interview/hashing/find-elements-which-are-present-in-first-array-and-not-in-second.htm
CC-MAIN-2021-49
refinedweb
710
70.63
Join the community to find out what other Atlassian users are discussing, debating and creating. Hi, we have the following issue with the letter font in a comment: If a user adds an attachment into a newly created comment, for example a Microsoft Outlook MSG file (message) this attachment is shown as a link. Then the user makes a bullet list with even this attachment line. When he then press ENTER (to the next bullet line) the text font will get smaller in the next line. Has anybody seen this already? Regards Tim Hi Tim, Sorry to hear about this problem. While I have seen other documented bugs around spacing in bullet lists in Jira, such as JRASERVER-33394, I have not yet found this specific bug documented that you describe here. I tried to recreate what you described in my 8.4.1 test instance of Jira, by adding a comment, and trying to insert both attachment links and external links into the first line of a bullet list. However I could not seem to replicate this problem here. I'd be interested to learn more about your environment here. Such as I'd be interested to figure out how I can reproduce this problem so that we can better document this behavior. Regards, Andy Hi @Andy_Heinzer , I am able to reproduce this issue in 8.3.4 and even in 8.4.2. Here is my step by step description of this issue: When I use the text mode there is a slight difference between the first and the second line. The second line is indented by a single space (?) character. When I copy and paste the whole text the indent dissapears: * [^Axians_R58700447910_Office365.PDF] * ^tes^ On my screenshot you will the the indent. The behavior is the same in Description. Copy and paste a URL (text) from the web browser does not change the font. Creating a external link using the editor menu seems to work well, too. HTH Regards Tim Hi Tim, Thanks for providing the detailed steps here. With these I was then also able to replicate this problem. I agree this is a bug in the way the visual editor in Jira server is working. It seems that it is incorrectly applying that caret ^ character here on that next line. In my research of this problem I came across a very similar reported bug in JRASERVER-64897. While the steps are a bit different here, I certainly think this is related to the problem you have found here. So I created a new bug that more clearly highlights the steps you took to locate this bug in JRASERVER-70097. I found that if you use a carriage return (enter) after creating the link, but before creating the bullet list / numbered list, that this formatting bug does not appear. Additionally this bug does not appear to extend itself to the text editor and seems to be bound specifically to the Visual editor. That said, thanks for raising this problem on Community. I am not sure when this bug might be fixed within Jira server, but you can watch that bug ticket for updates on this problem. I hope that one of these work-around above will be useful in the meantime. Cheers,.
https://community.atlassian.com/t5/Jira-questions/attachment-link-in-bullet-list-will-change-letter-font-in-next/qaq-p/1198955
CC-MAIN-2020-24
refinedweb
546
71.95
Python Programming, news on the Voidspace Python Projects and all things techie. IronPython & Windows Forms VI Note This article has moved. You can find the whole tutorial series at IronPython & Windows Forms. Like this post? Digg it or Del.icio.us it. Looking for a great tech job? Visit the Hidden Network Jobs Board. Posted by Fuzzyman on 2006-06-23 00:15:12 | | Categories: Python, Writing, IronPython The Build Cat Our full build (functional tests and unit tests) now takes around ten minutes. This is a problem. Our first take was to try and run the tests in a virtual machine. Our functional tests simulate user input with calls like SendKeys.SendWait. This is very unreliable with VMWare, spurious failures all over the shop. On Virtual PC it works, but is painfully slow. This isn't such a problem. Because we pair on everything, when we start a build we can simply go and work on the 'other' PC [1]. What's worse is when both pairs end up building simultaneously. Once one pair has checked in, the others have to do an update and a fresh build. To solve this we have introduced the build cat. If you want to check-in, you must acquire the Build cat before kicking off your build. We've called him Mutex. Oh, and welcome to our intern, Max. Hello Max. Like this post? Digg it or Del.icio.us it. Looking for a great tech job? Visit the Hidden Network Jobs Board. Posted by Fuzzyman on 2006-06-22 21:25:53 | | Categories: Python, Work, General Programming IronPython & Code Quality Tools At Resolver we would like to integrate some tools into our build process that check our code quality and test coverage. The most common coverage tool seems to be coverage.py, last worked on by Ned Batchelder. That, and all the others that I found use sys.settrace [1]. This works with Python stack frames and so isn't implemented for IronPython. The line number information is still in there, because we get it back in tracebacks. Currently there is no good way of getting coverage information with IronPython. coverage.py makes very simple use of sys.settrace. It uses the same function for global and local trace functions : def t(f, x, y): c[(f.f_code.co_filename, f.f_lineno)] = 1 return t Because IronPython does name mangling and dynamic code generation, it's not likely that native .NET tools would be much help. For code quality tools I looked at the three most common [2] : All three of these use compiler, compiler.ast, and/or parser. The parser module isn't available on IronPython, so none of these run. We can run the code quality checks over our code using CPython however. The basic information we want is stuff like : - Unused imports - Unused variables - Shadowing builtins Our tests ought to pick up errors like using undefined names (we're maintaining a steady 3:1 ratio of test code to production code), but it wouldn't hurt to be warned about these as well. I had to drop PyChecker straight away, it attempts to execute code to check it. I tested the other two with the following ridiculous snippet of code : x = 3 int = 6 def fred(hello): silly = 1 print y fred(10) james(20) PyFlakes reported the following errors (not enough !) : snippet.py:1: 'foobar' imported but unused snippet.py:6: undefined name 'y' snippet.py:9: undefined name 'james' It hasn't picked up on the builtin shadowing or the unused variables. PyLint gave back all we wanted, and lots more, so it looks like that is what we'll use. Here's the relevant part of the PyLint report : ************* Module snippet W: 3: Redefining built-in 'int' E: 6:fred: Undefined variable 'y' W: 5:fred: Unused variable 'silly' W: 4:fred: Unused argument 'hello' E: 9: Undefined variable 'james' W: 1: Unused import foobar Report ====== 8 statements analysed. When I reported these problems on the IP Mailing List the developers replied that they intended to implement sys.settrace soon, and would see how much work it would be to implement compiler (which requires parser). Cool response ! In fact there could be a pure Python alternative to the compiler module (parts of it anyway). EasyExtend has borrowed the parser from PyPy, and can generate a CST from Python source. PEP 269 (a pure Python version of pgen, which generates the Python parser from the grammar) has been bumped from Python 2.3 to 2.4, from 2.4 to 2.5, from 2.5 to 2.6... sigh Like this post? Digg it or Del.icio.us it. Looking for a great tech job? Visit the Hidden Network Jobs Board. Posted by Fuzzyman on 2006-06-22 20:46:52 | | Categories: Python, Work, Tools, IronPython Pythonwin & Movable Python A user asked on the Movable Python Mailing List asked if it is possible to use the Pythonwin IDE [1] with Movable Python. Pythonwin itself is usually launched from an executable called Pythonwin.exe. As you might imagine, this fails miserably when we use it with Movable Python, because it is looking for the installed version of Python. Thankfully most of Pythonwin is implemented in Python code, using the various win32 extension modules. The exe file is just a convenience stub. Mark Hammond came to the rescue and provided a pure Python file that does the same :() Download a copy of the pywin32 extensions. [2] Copy the pythonwin directory into your movpy\lib directory. Save the snippet of code above as pythonwin.py in the pythonwin directory you just created. If you run this file from Movable Python, the IDE will run. Like this post? Digg it or Del.icio.us it. Looking for a great tech job? Visit the Hidden Network Jobs Board. Posted by Fuzzyman on 2006-06-22 13:45:50 | | Categories: Python, Projects Visual Studio Tools for Office Today I went to Microsoft Office Developers Conference. It was a long event, but more interesting than I expected and a lot of information to take in. It was basically Microsoft selling the features of the Office 2007 suite to UK developers by demoing what can be done with it. Most of the 'impressive' features are enterprise level ones, that require setting up large infrastructures for document management across organisations. They demo well, but look very complex to use. Lots of other nice touches though. The thing that was most interesting, and even potentially relevant to Resolver, was VSTO. VSTO and its bedfellow VSTA allow you to write Windows Forms applications that are hosted within the office suite applications (and can access the current document and data). Microsoft are attempting to make Office a development platform, and hope that ISV will create applications that depend on office. This tool, currently in beta, can use any .NET language; yes that means IronPython. It particularly leverages custom task panes and the ribbon, both of which are new in Office 12 (i.e. Office 2007). This makes your application 'native' to office. You can embed controls within documents, but unlike previous techniques for working with Office, this is done on the application level rather than at the document level (read per document). You ship your application with a VSTO runtime (which probably needs to be licensed) and the user must have the relevant office application installed to use it. Despite the obvious cynical conclusions, it does open up new possibilities for interaction with the office applications. Given that your target market almost certainly has these tools installed and is familiar with their basic UI, it is a new deployment technique; a way into markets currently held in stranglehold by Microsoft. When version three of VSTO ships it will also work with the major Office 11 (2003) applications but not all of them. VSTA is intriguing. It is a version of Visual Studio that you license and ship with your VSTO application and allows your users to customise the application. Presumably you control how it can be customised. Not so relevant to us perhaps, but it certainly has possibilities... Like this post? Digg it or Del.icio.us it. Looking for a great tech job? Visit the Hidden Network Jobs Board. Posted by Fuzzyman on 2006-06-21 23:48:42 | | Categories: Work, General Programming Python Bits & Pieces Lots of good (and new) things have been happening in the Python world recently. Here's a quick overview of the few that have caught my eye. This is a safe pickle replacement [1]. This module allows you to safely save and restore (serialize) Python data structures including classes. With the help of psyco it can allegedly approach the performance of cPickle despite being written in around three hundred lines of Python. It is also compatible with the Pickle interface and protocols, so any investment you have put into making your objects Pickle compatible isn't wasted if you switch. Lawrence Oluyede - Summer of Code Lawrence has taken on a google Summer of Code project to port some of the CPython standard library modules to using ctypes. Now that (?) ctypes is supported in PyPy it means that more of the standard library will be available to it. You can read about his progress in his blog. The news from the PyPy project in recent months has been getting more and more exciting. It looks like it won't be long before it has practical uses. Anthony Baxter has announced the customary Python trunk freeze prior to the release of Python 2.5, beta 1. There is also an early version of py2exe available for Python 2.5. Now that I have returned to working on Movable Python I will try and release a Python 2.5 version ASAP. IronPython beta 8 is now available. It has lots of bugfixes that are helpful to us at Resolver Systems and it's good to see the team continue to make steady progress. I've nearly completed another entry in my series on IronPython & Windows Forms by the way. I'm aiming for one a week, and on average I think I'm still ahead of the game. The High Level Virtual Machine is a platform built on the Low Level Virtual Machine. It aims to provide a common and efficient runtime for implementing dynamic languages; including Python. You provide a language definition and production rules for turning source code into an HLVM AST, and it compiles the AST to bytecode [2] (Which the LLVM executes using it's JIT compiler). As well as being an ideal platform to experiment with languages on, it opens up the possibility of language interoperability (since all languages compile to the same AST). Unfortunately implementing Python seems to be quite far down the line in their plans. This either reflects the developers personal preferences, or the fact that the dynamic features of Python make it harder to implement. Either way I'm very tempted to experiment. I would like to write a compiler sometime, perhaps implementing my Python Extension Language idea. If I do, I will probably target the .NET platform and work my way through this excellent little online book "Let's Build a Compiler for the CLR". The difficult part (even for a simple language) would be static type annotation from source code. As well as Mono (as a cross platform target for a dotnet language), there is a GNU project called Portable dotnet that I've only recently come across. That's enough for today, except perhaps to say that tomorrow I'm off to the Microsoft Office System Developer Conference in London. I'm not sure I'll have anything interesting to report back here though. Hopefully next year I will be able to afford to go to one of the Python conferences. Like this post? Digg it or Del.icio.us it. Looking for a great tech job? Visit the Hidden Network Jobs Board. Posted by Fuzzyman on 2006-06-20 12:36:29 | | Categories: Python, IronPython PyZine Goes Free PyZine has long been an interesting and useful resource for Python programmers. Unfortunately it has languished recently, due to the difficulties the publishers have had in sourcing quality articles [1]. The publishers behind PyZine have their fingers in some other projects (like Open Source Experts) and until their attention returns to PyZine, they have opened up the archives. So (for the moment at least), you can freely browse the eight published online issues, which have a wealth of material and information. Like this post? Digg it or Del.icio.us it. Looking for a great tech job? Visit the Hidden Network Jobs Board. Posted by Fuzzyman on 2006-06-19 11:28:31 | | Categories: Python, Writing Pythonutils I haven't updated the Pythonutils Package for a while. Several of the modules in it have more recent versions available, so it's about time. As soon as StandOut 3 is ready for release, I will also release an updated Pythonutils package. I'm going to drop the listquote module, as it is unloved and unused [1]. I'm also considering dropping the pythonutils namespace and making pythonutils a bundle of modules rather than a package. For backwards compatibility in the forthcoming 0.3.0 release I can include a pythonutils.py module which imports everything and behaves like the pythonutils package. Like this post? Digg it or Del.icio.us it. Looking for a great tech job? Visit the Hidden Network Jobs Board. Posted by Fuzzyman on 2006-06-18 18:32:19 | | Categories: Projects, Python StandOut 3.0.0 Alpha I managed to go an entire week without a Python related entry. Of course, despite this lots of things have been going on. I have a stack of potential blog entries metaphorically at my side, some of which may never see the light of day. But first the most important thing. There is now a new version of StandOut in subversion : StandOut 3.0.0 Alpha There are also unit tests to go with it. StandOut is a simple class that redirects sys.stdout and sys.stderr. This serves two purposes. By passing in a file name when you instantiate StandOut it can log everything printed (and all errors) to a file. You can also assign messages a priority, and tell StandOut what priority of messages to actually display, making implementing different "verbosity levels" very easy. When I first wrote StandOut I thought I would use the verbosity features a lot. I didn't actually use it for anything except logging until recent changes in rest2web. I was amazed that it worked, and that it was easy to use. Unfortunately along the way, StandOut had grown lots of cruft and complicated ways of using it that nobody needed. This version is a complete rewrite. The code is better quality, and it has a simpler and straightforward API. It isn't compatible with the previous version, so programs currently using StandOut will need minor modifications to use the new one. Enough prattle, here's how to use it. By default StandOut diverts both stdout and stderr. If you pass in a filename (optional) then that will be used to log to. Output on stderr will be prefixed with "[err] " [1]. import sys from standout3 import StandOut filename = os.path.join(os.path.expanduser('~'), 'logfile.txt') standout = StandOut(filename) print 'hello' # replace with real code ;-) sys.stderr.write('This is an error') standout.close() The log file will then contain : hello [err] This is an error The standout.close() call restores sys.stdout and sys.stderr to their normal state and closes the logfile. If an unhandled exception occurs during your code, execution will terminate. If you want to guarantee that close is still called, you can put it inside a try:... finally: block. Unfortunately the exception will then be raised (by the finally clause) after close has been called. You can use the following trick to get round this : import sys import traceback from standout3 import StandOut filename = os.path.join(os.path.expanduser('~'), 'logfile.txt') standout = StandOut(filename) try: raise TypeError("Some Error") finally: traceback.print_last() standout.close() You will see that the error still appears in the logfile in the usual way. On the other hand, if you just let your program terminate (without trapping errors like this), then the open log file will be garbage collected and closed for you. By default all messages have a priority of five. There are four possible output methods : - The stdout stream (outStream) - Logging output to the file (outLogfile) - The stderr stream (errStream) - Logging errors to the file (errLogfile) You can configure their threshold independently, but there are easier ways as well. If a message has a equal or higher priority than the threshold for an output method, it will be passed on. If the priority is lower than the threshold, it will be dropped. Easy hey. The default threshold for the output stream and file is five. The default threshold for the error stream and file is zero (everything logged). import sys from standout3 import StandOut filename = os.path.join(os.path.expanduser('~'), 'logfile.txt') standout = StandOut(filename, priority=4) print 'Priority is ', standout.priority print 'Output message with a priority of four' print >> sys.stderr, 'Error message with a priority of four' standout.priority = 6 print 'Output message with a priority of six' print >> sys.stderr, 'Error message with a priority of six' standout.close() Unfortunately the message telling you that the priority is four won't be displayed because the threshold is above the priority. (Threshold 5, priority 4.) Logged to the file will be : [err] Error message with a priority of four Output message with a priority of six [err] Error message with a priority of six There is also another way of setting the priority of individual messages, which you may prefer. This is what I used to implement varying levels of verbosity in rest2web. import sys from standout3 import StandOut filename = os.path.join(os.path.expanduser('~'), 'logfile.txt') standout = StandOut(filename, errThreshold=5) sys.stdout.write('Message with a priority of six', 6) sys.stderr.write('Error with a priority of six', 6) sys.stdout.write('Message with a priority of four', 4) sys.stderr.write('Error with a priority of four', 4) standout.close() Because we explicitly set the threshold for errors to five, only output and errors with a priority of five (or greater) will be passed on. The above code results in the following output : Message with a priority of six [err] Error with a priority of six The following properties (which can also be passed in as keyword arguments) are used to set different thresholds for the different output methods : - standout.threshold - get or set the threshold for all of them - standout.outThreshold - get or set the threshold for the output stream and output logfile - standout.errThreshold - get or set the threshold for the error stream and error logfile - standout.outStreamThreshold - get or set the threshold for the output stream - standout.outLogfileThreshold - get or set the threshold for the output logfile - standout.errStreamThreshold - get or set the threshold for the error stream - standout.errLogfileThreshold - get or set the threshold for the error logfile If the output methods have different thresholds, then standout.threshold (etc) will return -1. There is currently no way of logging to a separate file for errors, other than instantiating StandOut twice : import sys from standout3 import StandOut filename = os.path.join(os.path.expanduser('~'), 'logfile.txt') errorFile = os.path.join(os.path.expanduser('~'), 'errorLog.txt') standout1 = StandOut(filename, stdErr=False) standout2 = StandOut(errorFile, stdOut=False) ... # code standout1.close() standout2.close() I developed this using the test driven development methodolgy we use at work. This guides the development process and results in better code and a better API. Someday I may do a blog entry about it. The result is an 8k module with 38k of tests. I'm sure the tests could be better factored, but I think the code is pretty good and well tested. There are still a few questions before I do a final release : - Should I build in a way to log to a separate error file with a single instance of StandOut ? - If you can specify a separate error file, should errPrefix default to '' ? - Should you be able to pass in open files as well as filenames ? I personally think that using two instances for separate logfiles is fine. This kind of answers questions one and two. I don't see the need for passing around open files, but it would make testing easier. Anyway, feedback appreciated. Like this post? Digg it or Del.icio.us it. Looking for a great tech job? Visit the Hidden Network Jobs Board. Posted by Fuzzyman on 2006-06-18 18:07:18 | | Categories: Python, Projects Archives This work is licensed under a Creative Commons Attribution-Share Alike 2.0 License. Counter...
http://www.voidspace.org.uk/python/weblog/arch_d7_2006_06_17.shtml
crawl-002
refinedweb
3,502
66.64
Written by Marcel Wiget For the first time in history, it is possible to run Junos instances on a Mac or PC laptop to learn and test Junos and its routing protocols. This blog post walks you through getting up and running using Docker Desktop, enabling ssh and applying basic Junos automation scripts via PyEz, all from the comfort of your laptop. Table of Contents Some Background cRPD Overview Requirements Load and Run cRPD SSH and NETCONF cRPD Access Persistent cRPD Configuration Connecting cRPD Instances Together Where To Go From Here? Junos has been the network operating systems on routers and switches for nearly 25 years now, tightly integrated on Juniper hardware. When I joined Juniper 12 years ago as a Systems Engineer, I heard about an unofficial project called “Olive”, which allowed Junos to run in a virtual machine. I used that unofficial code to get up to speed on Junos and the various routing protocols. But it was a bit like “walking on thin ice”, as some features just weren’t supported and there was nowhere to report them to. A few years ago, the vMX () was born, which separated the Junos control plane from the forwarding plane, each running in their separate virtual machines. I became an immediate fan and started to deploy them frequently on Linux servers. The resource requirements on memory, CPU and storage are, however, too big to run efficiently on laptops. Why not just take the control plane daemons from Junos (cli, mgd, rpd and the likes), port them to Linux, package it into a Linux container and ship it? It sounded really hard to me but engineering delivered the first commercial version with in 2019 with Junos 19.2! Containerized routing protocol process (cRPD) is Juniper’s routing protocol daemon (rpd) decoupled from Junos OS and packaged as a Docker container to run in Linux based environments. Wait, this blog talks about running cRPD on Docker Desktop (), available only for Mac (running OS/X) and PC (running Windows). Docker Desktop leverages native virtualization technique available on Windows, Hyper-V, and the Hypervisor framework on macOS, to run a small Linux VM, onto which containers can be deployed. Back to cRPD: rpd runs as a user space application and learns route state via its routing protocols and maintains it in the RIB (Routing Information Base) and downloads the routes into the FIB (forwarding information base) and shares it with the Linux kernel via netlink.(Source: cRPD Deployment Guide for Linux Server - cRPD on Linux Architecture) Two additional processes are required for cRPD to become fully functional: cli and mgd, which allow a user (or program) to manage the Junos configuration and retrieve state information. What about forwarding packets you might wonder by now. Well, that’s left to the Linux kernel (network namespace) the container runs in. The integration is amazingly seamless and opens up various use cases, starting from routing on the host to being the routing daemon in SONiC (). In fact, it runs on pretty much any netlink () supported Linux-based operating system. Now that you have the background knowledge, let’s get cRPD up and running on your Mac and PC based laptops. There are only 3 steps left to log into the cRPD CLI on your laptop: Figure 1: the same commands executed on Docker Desktop for PC and Mac. First, we need to load the cRPD image into Docker: Open a terminal window and execute the following command: The date shown next to the crpd:20.4R1.12 images show when the image was packaged and published by Juniper. Now we can launch cRPD as a daemon using a few command line options: And finally log into the CLI, by executing the cli process within the running container: Now go ahead and explore the Junos CLI a bit, try out various commands, enter config mode and exit. Version 20.4 and newer now support “show interfaces”, not just the “show interfaces routing”. Remember, cRPD handles neither packet forwarding nor interface configurations. If these commands give you an error like “error: the routing subsystem is not running”, then you likely haven’t used the “-ti” option for “docker run”. At least on: You may be tempted to check connectivity via “ping”. Well, bad luck. There is no ping in the cRPD CLI. Same for accessing the shell: No problem. You can access ping directly via bash shell in the container, by launching bash instead of cli via ‘docker exec’: Launching an application within the running container can come handy, e.g. if you just want to check a cli show command or try out Netconf interactively: Let’s try out Netconf via shell, first, find the rpc command for “show route” using the CLI, then execute the command in Netconf: Now let’s send this netconf command directly to the netconf application in cRPD. Note the use of option ‘-i’ instead of ‘-ti’. Because we pipe the command in, there is no TTY and avoid the error message: Can you reach the containers IP address from the host shell (PC or Mac)? Unfortunately, not on Mac and PC. Please note, we execute ‘cli show int’ without entering an interactive shell, so we don’t need the ‘-ti’ option. While it won’t hurt for this short output, it will interfere with output longer than 24 lines, because CLI invokes paging. Not specifying ‘-ti’ removes that issue: Ok, so no connectivity from the host to the container for PC and Mac. On Linux, this would have worked perfectly. Does this mean, usage of cRPD on Mac and PC is too limited and one needs to revert to installing and running a Linux VM (e.g. via Virtualbox, Parallels or VMware)? No. The workaround is to launch another container next to cRPD and execute the ping etc. from there. This will come in handy later, when trying out PyEz against cRPD. Done! Well, no, that’s just the start of it really. What about ssh and netconf access via PyEZ, what about persistent configuration storage and when do I need and how do I add a license key? Can I build a simple topology of interconnected cRPD instances? Let me address them all, one by one. First let’s stop and clean up the running container: The error shown above is fine, it just proves, we launched the container with automatic cleanup option (--rm). Starting with cRPD 20.4, ssh and Netconf can be configured like on any other Junos device via Junos configuration. Lets launch crpd1, configure root-authentication, create user lab and enable ssh with Netconf. I’m showing the committed config here. The password used is ‘lab123’: BTW using ‘\|’ allows the use of pipe between CLI commands to work just fine. We can’t just ssh or use Netconf from the host command shell, so we need to launch a container to do so. To find out the IP address of crpd1, we can use well known linux commands, executed in crpd1: Instead of using a plain alpine or ubuntu container, I’ll start with juniper/pyez, which gets downloaded automatically unless already present locally: Check to see if the ssh client is already installed. If not, find out what distribution this container is based on: In this case it’s Alpine. We can add ssh client with `apk add openssh`: Before launching ssh, find the local IP of the alpine container: The IP addresses of crpd1 and pyez container are in the same subnet. Let’s try out ssh from alpine to crpd1: bash-5.0# ssh lab@172.17.0.2 It worked! Without exiting the alpine container, let’s fire up Python and query crpd1 using a tiny script (create that file based on the output shown), then execute it: In case you run an older version of cRPD, port 830 might not respond. Try port 22 instead: Both methods work on version 20.4. So far, we successfully launched a cRPD container and access its CLI or use programmatic access via NETCONF. But once the container terminates, the applied configuration is lost. You could extract the configuration via “show config” or transfer it using scp. But there is better way. Simply mount a folder form the host into the cRPD container, where configuration changes are saved and remain available after an instance terminates. This is also a great way to populate a configuration right when an instance launches. Docker run has an option to mount volumes via its ‘—volume’ argument. A cRPD instance keeps its configuration within the container filesystem under /config. Here how this can be done, first on Mac, then on PC: And on PC (replacing $PWD with %CD%): Now any configuration change done within the crpd1 instance is saved persistently in the host folder crpd1: There is also a folder called ‘license’. This one will store license keys added via CLI or thru the Junos configuration. More often than not, a way to build point-to-point links between containers was needed in order to test functionalities requiring true L2 connectivity without a bridge in between, as typically provided by docker networking. A linux veth link generates 2 endpoint interfaces that can be moved into the containers network namespaces. On Mac and PC, it gets a bit trickier, as there is no easy access to the actual Linux VM running the containers. The trick is to launch a helper container to create the veth link and stitch them. That container requires access to the docker socket and share the PID namespace from the Linux kernel, which can all be done using docker run command line options. The “magic” of creating the links and adding them into a running container is done via a container marcelwiget/link-containers:latest, automatically built from this repo:, and published to hub.docker.com: On Mac: Once executed, you get 2 containers running, connected together with 2 links at eth1 and eth2 with their configurations saved automatically on the host (your laptop) in the folders ./crpd1 and ./crpd2, relative from your current directory you launched the containers from. So far, we launched just a single cRPD instance on a laptop, but what if you need a larger topology? All doable, but it requires some automation, ideally via docker-compose, which is automatically installed with Docker Desktop. A working example to build a mesh of many instances, that works on OSX (likely too on Windows), can be found here: (search for OSX in the README.md). A lot more information on cRPD can be found on Juniper Techpubs, e.g. in this cRPD Deployment Gguide for Linux: I hope you enjoyed this intro, if you find errors or have suggestions for improvements, please let us know below in the comment section.
https://community.juniper.net/answers/blogs/marcel-wiget1/2021/02/17/juniper-crpd-204-on-docker-desktop?CommunityKey=2f0a624d-2503-4ea5-8e6c-97cf5598792d
CC-MAIN-2021-43
refinedweb
1,804
59.64
How to use Python's enumerate and zip to iterate over two lists and their indices. enumerate- Iterate over indices and items of a list¶ The Python Cookbook (Recipe 4.4) describes how to iterate over items and indices in a list using enumerate. For example: alist = ['a1', 'a2', 'a3'] for i, a in enumerate(alist): print i, a Results: 0 a1 1 a2 2 a3 zip- Iterate over two lists in parallel¶ I previously wrote about using zip to iterate over two lists in parallel. Example: alist = ['a1', 'a2', 'a3'] blist = ['b1', 'b2', 'b3'] for a, b in zip(alist, blist): print a, b Results: a1 b1 a2 b2 a3 b3 enumerate with zip¶ Here is how to iterate over two lists and their indices using enumerate together with zip: alist = ['a1', 'a2', 'a3'] blist = ['b1', 'b2', 'b3'] for i, (a, b) in enumerate(zip(alist, blist)): print i, a, b Results: 0 a1 b1 1 a2 b2 2 a3 b3 Related posts - An example using Python's groupby and defaultdict to do the same task — posted 2014-10-09 - If you're working with last lists and/or memory is a concern, using the itertools module is an even better option. from itertools import izip, count alist = ['a1', 'a2', 'a3'] blist = ['b1', 'b2', 'b3'] for i, a, b in izip(count(), alist, blist): print i, a, b yields the exact same result as above, but is faster and uses less memory. >>> def foo(): ... for i, x, y in izip(count(), a, b): ... pass ... >>> def bar(): ... for i, (x, y) in enumerate(zip(a, b)): ... pass ... >>> delta(foo) 0.0213768482208 >>> delta(bar) 0.180979013443 where a = b = xrange(100000) and delta(f(x)) denotes the runtime in seconds of f(x). Jeremy, Thanks for the tip and the clear example and demonstration of the performance benefit. I had heard of itertools but have not really used it. It was great to talk to you today and I hope I can talk to you again soon. Thanks for the zip example, I grok it now. Jeremy, Thanks for the example, It is very helpful. I have set of n set, each with different number of elements I wanted to find all possible combinations of n elements, one from each set. Consider two sets (e1,e2,e3) (e4,e5) output required is as follows (e1,e4) (e1,e5) (e2,e4) (e2,e5) (e3,e4) (e3,e5) I do not know the number of such sets in advance. Nitin: In order to use zip to iterate over two lists - Do the two lists have to be the same size? What happens if the sizes are unequal? Thanks. Thx man helped me alot nice example btw re:#8, unequal list length: the result is truncated to the shorter list. See below for a discussion of how to use the longest list instead: short answer for py2.6+: use "map(None, alist, blist)" dunno what the equivalent is in py3+ when iterating through unequal length lists using itertools import itertools a1=[1,2,3,6,7,9] c1=['a','a','c','d'] b1=[10,20,30,40,50,60] d1=[11,12,13,14,15,16,17,18] mylist = list(itertools.izip_longest(a1,b1,c1,d1)) a1=[1,2,3,6,7,9] for items in mylist: litems=list(items) if items[0] is not None and items[1] is not None: a_old = items[0] b_old = items[1] if items[0] is None and items[1] is None: litems[0]= a_old litems[1]= b_old a,b,c,d=litems print a,b,c,d is ther any other better way to give previous value if None occurs for any field. disqus:2412310580 Very useful page with clear examples, thanks. disqus:3273150118
https://www.saltycrane.com/blog/2008/04/how-to-use-pythons-enumerate-and-zip-to/
CC-MAIN-2021-17
refinedweb
627
65.66
i used Microsoft Visual C++ for the resource script and everything is fine, i made a menu, an icon, a cursor, and version info and saved it all, brought it home (i made it on the schools compiler) and plugged the Resource.h and Resource.rc into my Bloodshed Dev C++ project, and created my window. now the window itsself doesn;t have any errors at all. but how do i include the custom cursor i made and icon? WndClass.hCursor = LoadCursor (NULL, IDC_MAINCUR); WndClass.hIcon = LoadIcon (NULL, IDI_ICON1); thats how i thought and it wont work. but heres another thing: theres a problem in the resource script. its trying to: #include "afxres.h" and it sais that there is no such file or directory. i set the settings of my compiler to Win32GUI. but the frig is going on here!
http://cboard.cprogramming.com/windows-programming/30170-wtf-wrong-resource-script.html
CC-MAIN-2014-10
refinedweb
141
76.93
I would like to have a dataset for making withdrawals and i would like every user to enter the amount to withdraw depending on the balance on a field. If the user inputs a value more than the balance return an error message.If the user inputs a value less than the balance, it should subtract that value and set the new balance. Anyone with an idea how this can be done with a pseudo code, I will truly appreciate. I would start with creating a collection with columns of the user ID and the amount he currently have. Once the user clicks on a button to perform an action, I would query the collection to find the user in the collection. Once I find the relevant record, I would check whether the current amount is above the amount added (using if statement). If it's above, you can use the update function to update the record with the new value. Otherwise, you can show an error message. I hope it's clear. Good luck, Tal. Thanks for your response. The problem comes when the the input value needs to be subtracted from the from the current amount to get the new amount and the Update function. Here is my code. export function button1_click(event, $w) { let val = $w("#input9").value; console.log(val); let item = $w("#dataset2").getCurrentItem(); console.log(item); if( val > item.field ) { // I would like the to subtract the input8 from input9 and update input9 as the new accountbalance } } I will appreciate if you will help modify the code. Hi, getCurrentItem function here is not relevant. You should query the collection and search for the userId. After getting the relevant record, you should check: I recommend re-reading the answer above and use the documentation links as they have examples of how to use those functions. Note that the code depends on how you've set your collection (the fields names). Good luck, Tal. Thanks Tal. I will try the code using the fiels i have on my db. Tal, the code above does not subtract the two input8 as expected. Here is the full code. import wixData from 'wix-data'; export function button300_click(event, $w) { let val = $w("#input9").value; console.log(val); let item = $w("#dataset5").getCurrentItem(); console.log(item); if( val < item.field ) { // I would like the to subtract the input8 from input9 and update input9 as the new accountbalance } if (val < item.accB ){ let newaccB = item.accB - val ; let toUpdate = { "_id": item._id , "accB": newaccB , "userId": item.userId }; wixData.update("custdetails", toUpdate) ; } }
https://www.wix.com/corvid/forum/community-discussion/performing-calculation-on-a-field-depending-on-user-input
CC-MAIN-2020-05
refinedweb
430
76.11
Opened 4 years ago Closed 2 years ago Last modified 2 years ago #20889 closed Bug (fixed) HttpResponseBase._convert_to_charset complains about newline it inserted itself Description Consider the following snippet: # coding=utf-8 from django.http import HttpResponse h = HttpResponse() f = 'zzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzz a\xcc\x88'.decode('utf-8') h['Content-Disposition'] = u'attachment; filename="%s"' % f This contains a "OSX-Style" Umlaut (a + ¨), which triggers the "convert to mime encoding" path of _convert_to_charset. However, this string is so long that email.header.Header, which is used to mime-encode, will split the line. Later _convert_to_charset explodes because the header now contains a newline. $ ./manage.py shell Python 2.7.5 (default, Aug 1 2013, 01:01:17) [GCC 4.2.1 Compatible Apple Clang 4.1 ((tags/Apple/clang-421.11.66))] on darwin Type "help", "copyright", "credits" or "license" for more information. (InteractiveConsole) >>> import ttt Traceback (most recent call last): File "<console>", line 1, in <module> File "/Users/mjl/Projects/cms_sites/cc.intranet/cms/ttt.py", line 5, in <module> h['Content-Disposition'] = u'attachment; filename="%s"' % f File "/Users/mjl/Projects/cms_sites/cc.intranet/lib/python2.7/site-packages/django/http/response.py", line 110, in __setitem__ value = self._convert_to_charset(value, 'latin1', mime_encode=True) File "/Users/mjl/Projects/cms_sites/cc.intranet/lib/python2.7/site-packages/django/http/response.py", line 105, in _convert_to_charset raise BadHeaderError("Header values can't contain newlines (got %r)" % value) BadHeaderError: Header values can't contain newlines (got '=?utf-8?q?attachment=3B_filename=3D=22zzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzz_a?=\n =?utf-8?b?zIgi?=') Solution: django/http/response.py#L163 should probably use Header(value, 'utf-8', maxlinelen=1000) to avoid generating continuation lines. Agreed, the "1000" is somewhat arbitrary, but realistically this fixes the problem. Change History (17) comment:1 Changed 3 years ago by comment:2 Changed 3 years ago by comment:3 Changed 3 years ago by comment:4 Changed 3 years ago by comment:5 Changed 3 years ago by comment:6 Changed 3 years ago by comment:7 Changed 3 years ago by The bug you are encountering poly, is (I think) because of a bug in the Python standard library that has never been solved. The bug was reported over here: However, it was closed for some reason. I can still reproduce the bug with this snippet: from email.header import Header import sys=sys.maxsize ).encode() which output is: ==?= So it's because there's one character in the header that can't be encoded to latin-1 which raises a UnicodeError, which will be catched and will be encoded using the Header class from the python standard library in So this happens over here: Well, it correctly gives the maxlinelen. Now, what goes wrong in there? If you dive in to the code, I tackled it down to the _encode_chunks the Header class. This method takes the maxlinelen argument (which is ultimately based on the value initially provided in the init method in the code I posted above). However, it doesn't give it to the header_encode method that is used in this piece of code in this _encode_chunks method: if charset is None or charset.header_encoding is None: s = header else: s = charset.header_encode(header) So if charset.header_encoding is set it uses this method to encode the header, but then the maxlinelen is lost. This is where it goes wrong. I'll keep this ticket assigned to me as I'm probably going on with it tomorrow. comment:8 Changed 3 years ago by I just created a ticket in the Python bugtracker about this: comment:9 Changed 3 years ago by comment:10 Changed 3 years ago by Actually, I'm not sure that the check for newline is actually correct. There is no problem with a newline in a header as long as it's followed by white space, which means it's a continuation line and will be folded back at the other end. I understand where that check comes from though, there have been too many header injections out there. comment:11 Changed 3 years ago by Is there any harm in leaving the original fix in 1.7? Should we leave this ticket open or is it something that should by solved in Python itself? comment:12 Changed 3 years ago by Closing as needsinfo as it's not clear to me what we can do in Django to solve this issue and I haven't been able to get anyway to follow up in 3 months. comment:13 Changed 2 years ago by comment:14 Changed 2 years ago by I've encountered this bug on a production site. I'm suggesting to check for newlines in user-provided values, before formatting the header through Python API to workaround this issue. In 6dca603abb0eb164ba87657caf5cc65bca449719:
https://code.djangoproject.com/ticket/20889?cversion=0&cnum_hist=7
CC-MAIN-2017-09
refinedweb
802
54.83
This document is also available in these non-normative formats: XML and 2004-02-27 Diff. Copyright © 2004 drafted by the editor for discussion by the TAG and has no normative status at this Element and Attribute Names 4 QNames in Other XML Names 4.1 QNames in Other Specifications 4.2 Namespace Bindings 5 Architectural Observations 6 Architectural Statement 7 References A Use Case: XML Canonicalization (Non-Normative).. Using a QName as a shortcut for a {URI, local-name} pair is often convenient, but it carries a price. In order to identify QNames in content, a processor must understand the syntax and possibly the semantics of the content. The “ x:p” in the preceding example can only be recognized as a QName by a processor that understands both XPath (in order to parse the attribute value) and XSLT (in order to know which attributes contain XPath expressions)., therefore accept that there are some applications which use in-scope namespaces and some which use their own mechanisms. However, the namespace binding framework defined by the [XML Namespaces] is well established and widely supported. The cost associated with defining and deploying an alternate mechanism is very large and should be avoided wherever possible. Because there is some possibility of variation in the way namespace bindings are established, even if a QName can be identified in content, it may be difficult or impossible to determine what {URI, local-name} it represents. The mapping may depend. use a QName for identification when a URI would serve. That said, the TAG recognizes that there are sometimes pragmatic reasons for chosing short, lexical representations of more complex names and accepts that QNames are an established mechanism for doing so. Further, it must be observed that some things are identified by QNames: element and attribute names, types in W3C XML Schema, etc. Where there is a compelling reason to use QNames instead of URIs for identification, it is imperative that specifications provide a mapping between QNames and URIs, if such a mapping is possible. Finally, we observe that a whole class of interpretation problems can be avoided if the use of QNames can be restricted to contexts where their identification is natural and unambiguous (element and attribute names, simple content of type xs:QName, etc.) and we encourage developers to employ such restrictions wherever possible..
http://www.w3.org/2001/tag/doc/qnameids-2004-02-27.html
CC-MAIN-2016-44
refinedweb
389
50.16
import "github.com/hugelgupf/p9/p9" Package p9 is a 9P2000.L implementation. Servers implement Attacher and File interfaces. Clients can use Client. buffer.go client.go client_file.go file.go handlers.go messages.go p9.go path_tree.go pool.go server.go transport.go version.go const ( // NoUID is a sentinel used to indicate no valid UID. NoUID UID = math.MaxUint32 // NoGID is a sentinel used to indicate no valid GID. NoGID GID = math.MaxUint32 ) var AttrMaskAll = AttrMask{ Mode: true, NLink: true, UID: true, GID: true, RDev: true, ATime: true, MTime: true, CTime: true, INo: true, Size: true, Blocks: true, BTime: true, Gen: true, DataVersion: true, } AttrMaskAll is an AttrMask with all fields masked. Debug can be assigned to log.Printf to print messages received and sent. ErrBadVersionString indicates that the version string is malformed or unsupported. ErrNoValidMessage indicates no valid message could be decoded. ErrOutOfFIDs indicates no more FIDs are available. ErrOutOfTags indicates no tags are available. ErrUnexpectedTag indicates a response with an unexpected tag was received. ErrVersionsExhausted indicates that all versions to negotiate have been exhausted. CanOpen returns whether this file open can be opened, read and written to. This includes everything except symlinks and sockets. HighestVersionString returns the highest possible version string that a client may request or a server may support. StatToAttr converts a Linux syscall stat structure to an Attr. VersionSupportsMultiUser returns true if version v supports multi-user fake directory permissions and ID values. type Attacher interface { // Attach returns a new File. // // The client-side attach will be translate to a series of walks from // the file returned by this Attach call. Attach() (File, error) } Attacher is provided by the server. type Attr struct { Mode FileMode UID UID GID GID NLink NLink RDev Dev Size uint64 BlockSize uint64 Blocks uint64 ATimeSeconds uint64 ATimeNanoSeconds uint64 MTimeSeconds uint64 MTimeNanoSeconds uint64 CTimeSeconds uint64 CTimeNanoSeconds uint64 BTimeSeconds uint64 BTimeNanoSeconds uint64 Gen uint64 DataVersion uint64 } Attr is a set of attributes for getattr. func (a *Attr) Apply(mask SetAttrMask, attr SetAttr) Apply applies this to the given Attr. String implements fmt.Stringer. type AttrMask struct { Mode bool NLink bool UID bool GID bool RDev bool ATime bool MTime bool CTime bool INo bool Size bool Blocks bool BTime bool Gen bool DataVersion bool } AttrMask is a mask of attributes for getattr. Contains returns true if a contains all of the attributes masked as b. Empty returns true if no fields are masked. String implements fmt.Stringer. Client is at least a 9P2000.L client. NewClient creates a new client. It performs a Tversion exchange with the server to assert that messageSize is ok to use. You should not use the same conn for multiple clients. Attach attaches to a server. Note that authentication is not currently supported. Close closes the underlying connection. Version returns the negotiated 9P2000.L.Google version number. ClientOpt enables optional client configuration. WithClientLogger overrides the default logger for the client. WithMessageSize overrides the default message size. ConnError is returned in cases of a connection issue. This may be treated differently than other errors. Is reports whether any error in err's chain matches target. DefaultWalkGetAttr implements File.WalkGetAttr to return ENOSYS for server-side Files. WalkGetAttr implements File.WalkGetAttr. Dev is the device number of an fs object. While this type has no utilities, it is useful in order to force linux+amd64 only developers to cast to Dev for the Dev field, which will make their code compatible with other GOARCH and GOOS values. type Dirent struct { // QID is the entry QID. QID QID // Offset is the offset in the directory. // // This will be communicated back the original caller. Offset uint64 // Type is the 9P type. Type QIDType // Name is the name of the entry (i.e. basename). Name string } Dirent represents a directory entry in File.Readdir. String implements fmt.Stringer. Dirents is a collection of directory entries. Find returns a Dirent with the given name if it exists, or nil. ErrBadResponse indicates the response didn't match the request. func (e *ErrBadResponse) Error() string Error returns a highly descriptive error. ErrInvalidMsgType is returned when an unsupported message type is found. func (e *ErrInvalidMsgType) Error() string Error returns a useful string. ErrMessageTooLarge indicates the size was larger than reasonable. func (e *ErrMessageTooLarge) Error() string Error returns a sensible error. type FSStat struct { // Type is the filesystem type. Type uint32 // BlockSize is the blocksize. BlockSize uint32 // Blocks is the number of blocks. Blocks uint64 // BlocksFree is the number of free blocks. BlocksFree uint64 // BlocksAvailable is the number of blocks *available*. BlocksAvailable uint64 // Files is the number of files available. Files uint64 // FilesFree is the number of free file nodes. FilesFree uint64 // FSID is the filesystem ID. FSID uint64 // NameLength is the maximum name length. NameLength uint32 } FSStat is used by statfs. type File interface { // Walk walks to the path components given in names. // // Walk returns QIDs in the same order that the names were passed in. // // An empty list of arguments should return a copy of the current file. // // On the server, Walk has a read concurrency guarantee. Walk(names []string) ([]QID, File, error) // WalkGetAttr walks to the next file and returns its maximal set of // attributes. // // Server-side p9.Files may return linux.ENOSYS to indicate that Walk // and GetAttr should be used separately to satisfy this request. // // On the server, WalkGetAttr has a read concurrency guarantee. WalkGetAttr([]string) ([]QID, File, AttrMask, Attr, error) // StatFS returns information about the file system associated with // this file. // // On the server, StatFS has no concurrency guarantee. StatFS() (FSStat, error) // GetAttr returns attributes of this node. // // On the server, GetAttr has a read concurrency guarantee. GetAttr(req AttrMask) (QID, AttrMask, Attr, error) // SetAttr sets attributes on this node. // // On the server, SetAttr has a write concurrency guarantee. SetAttr(valid SetAttrMask, attr SetAttr) error // Close is called when all references are dropped on the server side, // and Close should be called by the client to drop all references. // // For server-side implementations of Close, the error is ignored. // // Close must be called even when Open has not been called. // // On the server, Close has no concurrency guarantee. Close() error // Open must be called prior to using ReadAt, WriteAt, or Readdir. Once // Open is called, some operations, such as Walk, will no longer work. // // On the client, Open should be called only once. The fd return is // optional, and may be nil. // // On the server, Open has a read concurrency guarantee. Open is // guaranteed to be called only once. // // N.B. The server must resolve any lazy paths when open is called. // After this point, read and write may be called on files with no // deletion check, so resolving in the data path is not viable. Open(mode OpenFlags) (QID, uint32, error) // ReadAt reads from this file. Open must be called first. // // This may return io.EOF in addition to linux.Errno values. // // On the server, ReadAt has a read concurrency guarantee. See Open for // additional requirements regarding lazy path resolution. ReadAt(p []byte, offset int64) (int, error) // WriteAt writes to this file. Open must be called first. // // This may return io.EOF in addition to linux.Errno values. // // On the server, WriteAt has a read concurrency guarantee. See Open // for additional requirements regarding lazy path resolution. WriteAt(p []byte, offset int64) (int, error) // FSync syncs this node. Open must be called first. // // On the server, FSync has a read concurrency guarantee. FSync() error // Create creates a new regular file and opens it according to the // flags given. This file is already Open. // // N.B. On the client, the returned file is a reference to the current // file, which now represents the created file. This is not the case on // the server. These semantics are very subtle and can easily lead to // bugs, but are a consequence of the 9P create operation. // // On the server, Create has a write concurrency guarantee. Create(name string, flags OpenFlags, permissions FileMode, uid UID, gid GID) (File, QID, uint32, error) // Mkdir creates a subdirectory. // // On the server, Mkdir has a write concurrency guarantee. Mkdir(name string, permissions FileMode, uid UID, gid GID) (QID, error) // Symlink makes a new symbolic link. // // On the server, Symlink has a write concurrency guarantee. Symlink(oldName string, newName string, uid UID, gid GID) (QID, error) // Link makes a new hard link. // // On the server, Link has a write concurrency guarantee. Link(target File, newName string) error // Mknod makes a new device node. // // On the server, Mknod has a write concurrency guarantee. Mknod(name string, mode FileMode, major uint32, minor uint32, uid UID, gid GID) (QID, error) // Rename renames the file. // // Rename will never be called on the server, and RenameAt will always // be used instead. Rename(newDir File, newName string) error // RenameAt renames a given file to a new name in a potentially new // directory. // // oldName must be a name relative to this file, which must be a // directory. newName is a name relative to newDir. // // On the server, RenameAt has a global concurrency guarantee. RenameAt(oldName string, newDir File, newName string) error // UnlinkAt the given named file. // // name must be a file relative to this directory. // // Flags are implementation-specific (e.g. O_DIRECTORY), but are // generally Linux unlinkat(2) flags. // // On the server, UnlinkAt has a write concurrency guarantee. UnlinkAt(name string, flags uint32) error // Readdir reads directory entries. // // offset is the entry offset, and count the number of entries to // return. // // This may return io.EOF in addition to linux.Errno values. // // On the server, Readdir has a read concurrency guarantee. Readdir(offset uint64, count uint32) (Dirents, error) // Readlink reads the link target. // // On the server, Readlink has a read concurrency guarantee. Readlink() (string, error) // Renamed is called when this node is renamed. // // This may not fail. The file will hold a reference to its parent // within the p9 package, and is therefore safe to use for the lifetime // of this File (until Close is called). // // This method should not be called by clients, who should use the // relevant Rename methods. (Although the method will be a no-op.) // // On the server, Renamed has a global concurrency guarantee. Renamed(newDir File, newName string) } File is a set of operations corresponding to a single node. Note that on the server side, the server logic places constraints on concurrent operations to make things easier. This may reduce the need for complex, error-prone locking and logic in the backend. These are documented for each method. There are three different types of guarantees provided: none: There is no concurrency guarantee. The method may be invoked concurrently with any other method on any other file. read: The method is guaranteed to be exclusive of any write or global operation that is mutating the state of the directory tree starting at this node. For example, this means creating new files, symlinks, directories or renaming a directory entry (or renaming in to this target), but the method may be called concurrently with other read methods. write: The method is guaranteed to be exclusive of any read, write or global operation that is mutating the state of the directory tree starting at this node, as described in read above. There may however, be other write operations executing concurrently on other components in the directory tree. global: The method is guaranteed to be exclusive of any read, write or global operation. FileMode are flags corresponding to file modes. These correspond to bits sent over the wire. These also correspond to mode_t bits. const ( // FileModeMask is a mask of all the file mode bits of FileMode. FileModeMask FileMode = 0170000 // ModeSocket is an (unused) mode bit for a socket. ModeSocket FileMode = 0140000 // ModeSymlink is a mode bit for a symlink. ModeSymlink FileMode = 0120000 // ModeRegular is a mode bit for regular files. ModeRegular FileMode = 0100000 // ModeBlockDevice is a mode bit for block devices. ModeBlockDevice FileMode = 060000 // ModeDirectory is a mode bit for directories. ModeDirectory FileMode = 040000 // ModeCharacterDevice is a mode bit for a character device. ModeCharacterDevice FileMode = 020000 // ModeNamedPipe is a mode bit for a named pipe. ModeNamedPipe FileMode = 010000 // Read is a mode bit indicating read permission. Read FileMode = 04 // Write is a mode bit indicating write permission. Write FileMode = 02 // Exec is a mode bit indicating exec permission. Exec FileMode = 01 // AllPermissions is a mask with rwx bits set for user, group and others. AllPermissions FileMode = 0777 // Sticky is a mode bit indicating sticky directories. Sticky FileMode = 01000 ) ModeFromOS returns a FileMode from an os.FileMode. FileType returns the file mode without the permission bits. IsBlockDevice returns true if m represents a character device. IsCharacterDevice returns true if m represents a character device. IsDir returns true if m represents a directory. IsExecutable returns true if m represents a file that can be executed. IsNamedPipe returns true if m represents a named pipe. IsReadable returns true if m represents a file that can be read. IsRegular returns true if m is a regular file. IsSocket returns true if m represents a socket. IsSymlink returns true if m represents a symlink. IsWritable returns true if m represents a file that can be written to. OSMode converts a p9.FileMode to an os.FileMode. Permissions returns just the permission bits of the mode. QIDType is the most significant byte of the FileMode word, to be used as the Type field of p9.QID. Writable returns the mode with write bits added. GID represents a group ID. Ok returns true if gid is not NoGID. NLink is the number of links to this fs object. While this type has no utilities, it is useful in order to force linux+amd64 only developers to cast to NLink for the NLink field, which will make their code compatible with other GOARCH and GOOS values. OpenFlags is the mode passed to Open and Create operations. These correspond to bits sent over the wire. const ( // ReadOnly is a Topen and Tcreate flag indicating read-only mode. ReadOnly OpenFlags = 0 // WriteOnly is a Topen and Tcreate flag indicating write-only mode. WriteOnly OpenFlags = 1 // ReadWrite is a Topen flag indicates read-write mode. ReadWrite OpenFlags = 2 // OpenFlagsModeMask is a mask of valid OpenFlags mode bits. OpenFlagsModeMask OpenFlags = 3 ) Mode returns only the open mode (read-only, read-write, or write-only). OSFlags converts a p9.OpenFlags to an int compatible with open(2). String implements fmt.Stringer. type QID struct { // Type is the highest order byte of the file mode. Type QIDType // Version is an arbitrary server version number. Version uint32 // Path is a unique server identifier for this path (e.g. inode). Path uint64 } QID is a unique file identifier. This may be embedded in other requests and responses. String implements fmt.Stringer. QIDGenerator is a simple generator for QIDs that atomically increments Path values. func (q *QIDGenerator) Get(t QIDType) QID Get returns a new 9P unique ID with a unique Path given a QID type. While the 9P spec allows Version to be incremented every time the file is modified, we currently do not use the Version member for anything. Hence, it is set to 0. QIDType represents the file type for QIDs. QIDType corresponds to the high 8 bits of a Plan 9 file mode. const ( // TypeDir represents a directory type. TypeDir QIDType = 0x80 // TypeAppendOnly represents an append only file. TypeAppendOnly QIDType = 0x40 // TypeExclusive represents an exclusive-use file. TypeExclusive QIDType = 0x20 // TypeMount represents a mounted channel. TypeMount QIDType = 0x10 // TypeAuth represents an authentication file. TypeAuth QIDType = 0x08 // TypeTemporary represents a temporary file. TypeTemporary QIDType = 0x04 // TypeSymlink represents a symlink. TypeSymlink QIDType = 0x02 // TypeLink represents a hard link. TypeLink QIDType = 0x01 // TypeRegular represents a regular file. TypeRegular QIDType = 0x00 ) Server is a 9p2000.L server. NewServer returns a new server. func (s *Server) Handle(t io.ReadCloser, r io.WriteCloser) error Handle handles a single connection. Serve handles requests from the bound socket. The passed serverSocket _must_ be created in packet mode. ServerOpt is an optional config for a new server. WithServerLogger overrides the default logger for the server. type SetAttr struct { Permissions FileMode UID UID GID GID Size uint64 ATimeSeconds uint64 ATimeNanoSeconds uint64 MTimeSeconds uint64 MTimeNanoSeconds uint64 } SetAttr specifies a set of attributes for a setattr. String implements fmt.Stringer. type SetAttrMask struct { Permissions bool UID bool GID bool Size bool ATime bool MTime bool CTime bool ATimeNotSystemTime bool MTimeNotSystemTime bool } SetAttrMask specifies a valid mask for setattr. func (s SetAttrMask) Empty() bool Empty returns true if no fields are masked. func (s SetAttrMask) IsSubsetOf(m SetAttrMask) bool IsSubsetOf returns whether s is a subset of m. func (s SetAttrMask) String() string String implements fmt.Stringer. UID represents a user ID. Ok returns true if uid is not NoUID. Package p9 imports 20 packages (graph) and is imported by 10 packages. Updated 2020-07-28. Refresh now. Tools for package owners.
https://godoc.org/github.com/hugelgupf/p9/p9
CC-MAIN-2020-34
refinedweb
2,793
60.61
Mercurial > dropbear view libtommath/bn_mp_clamp.c @ 475:52a644e7b8e1 pubkey-options * Patch from Frédéric Moulins adding options to authorized_keys. Needs review. line source #include <tommath.h> #ifdef BN_MP_CLAMP], */ /* trim unused digits * * This is used to ensure that leading zero digits are * trimed and the leading "used" digit will be non-zero * Typically very fast. Also fixes the sign if there * are no more leading digits */ void mp_clamp (mp_int * a) { /* decrease used while the most significant digit is * zero. */ while (a->used > 0 && a->dp[a->used - 1] == 0) { --(a->used); } /* reset the sign flag if used == 0 */ if (a->used == 0) { a->sign = MP_ZPOS; } } #endif /* $Source: /cvs/libtom/libtommath/bn_mp_clamp.c,v $ */ /* $Revision: 1.3 $ */ /* $Date: 2006/03/31 14:18:44 $ */
https://hg.ucc.asn.au/dropbear/file/52a644e7b8e1/libtommath/bn_mp_clamp.c
CC-MAIN-2022-21
refinedweb
121
58.99
HTTP dynamic streaming without FMS ?daslicht Nov 3, 2010 11:37 AM Hi, is it possible to use HTTP Streaming without a Flash Media Server ? Cheers Marc 1. Re: HTTP dynamic streaming without FMS ?Flashing Mathur Nov 3, 2010 11:48 AM (in response to daslicht) Hi, The answer is yes, but the only part missing in this is that you will not be able to showcase live videos. As you need to publish them to the fms which makes the incoming stream http ready and then lets you subscribe a stream through http. Okay so to start with http streaming you need to download f4fpackager and Http Origin module. To start with follow the link -7ffc.html for apache installation,- 7ffc.html for fragmentation and using f4f packager. 2. Re: HTTP dynamic streaming without FMS ?daslicht Nov 3, 2010 12:40 PM (in response to Flashing Mathur) Sounds great ! I try it. So basically I just need audio streaming and as far as I see, my contnet is proteced from downloading(since its fragmented), no ? 3. Re: HTTP dynamic streaming without FMS ?Flashing Mathur Nov 4, 2010 1:21 AM (in response to daslicht) Hi, I got your requirement but was unable to understand " my contnet is proteced from downloading(since its fragmented), no ?" Can you please elaborate more upon it. 4. Re: HTTP dynamic streaming without FMS ?daslicht Nov 4, 2010 4:16 AM (in response to Flashing Mathur) Thank you for your time ! Is it true that after segmenting the content, it is less easy to steal it, since there is no way to glue the segments together again ? I hope this sentence makes more sence, sorry Best Regards Marc 5. Re: HTTP dynamic streaming without FMS ?daslicht Nov 4, 2010 4:42 AM (in response to Flashing Mathur) Hello, ok I have now installed the Origin Module for Apache. When I try to use the packager, I get the following error: Error 12: Standard Exception: bad allocation I used the following settings: <offline> <input-file> Konstantin___Miko__the_Magic_Violin___Antaris_2009.flv </input-file> <output-path> Konstantin___Miko__the_Magic_Violin___Antaris_2009 </output-path> <fragment-duration> 25 </fragment-duration> <segment-duration> 25 </segment-duration> </offline> Any idea whats wrong ? 6. Re: HTTP dynamic streaming without FMS ?daslicht Nov 4, 2010 5:39 AM (in response to daslicht) OK , I did it ! I event get the audio playing with the flex spark videoplayer component. When I try it the following it is not working : <?xml version="1.0" encoding="utf-8"?> <s:Application xmlns:fx="" xmlns:s="library://ns.adobe.com/flex/spark" xmlns: <fx:Script> <![CDATA[ import flash.display.Sprite; import flash.media.Video; import flash.net.NetConnection; import flash.net.NetStream; import mx.events.FlexEvent; private var nc:NetConnection; private var ns:NetStream; //private var vid:Video; private var client:Object; public function SuperSimpleFLVPlayer ():void { // Initialize net stream nc = new NetConnection(); nc.connect (null); // Not using a media server. ns = new NetStream(nc); client = new Object(); ns.client = client; client.onMetaData = onMetaData_handler; client.onCuePoint = onCuePoint_handler // Play video //vid.attachNetStream ( ns );September Drive by Lightrocker_1.flv ns.play ( '' ); //ns.seek( 555.664 ); } //MetaData private function onMetaData_handler ( mdata:Object ):void { trace (mdata.duration); } private function onCuePoint_handler( item:Object ):void { trace("cuePoint"); trace(item.name + "\t" + item.time); } ]]> </fx:Script> </s:Application> 7. Re: HTTP dynamic streaming without FMS ?daslicht Dec 9, 2010 6:22 AM (in response to daslicht) noone ? 8. Re: HTTP dynamic streaming without FMS ?Flashing Mathur Dec 12, 2010 4:25 AM (in response to daslicht) Hey The latest queries you have posted are more related to the player side rather than what Http apache module does or f4fpackager does. Some links which may be of some help for you one suggestion would be for further any player related queries. Kindly post them at 9. Re: HTTP dynamic streaming without FMS ?daslicht Jan 3, 2011 6:12 AM (in response to Flashing Mathur) Thank you, I have successfully built a player with the needed functions now :
https://forums.adobe.com/message/3250581
CC-MAIN-2016-36
refinedweb
662
60.51
t/h stone crusher in canada t/h stone crusher in canada import 10 1000 t/h...limestone crusher quarries in canada,sell stone crusher capacity of 200 t h, Home > Product >sell stone crusher capacity of 200 t h. sell stone crusher capacity of 200 t h. 2020-08-08T05:08, ...Stone Crusher Made In Canada, Stone Crusher Made ...120 tph stone crusher in canada, 120 tph stone crusher in canada,home jaw rock crusher for sale in canada free jaw stone crushers salejaw crusher 120 t h price used jaw crushers plan 120 t h. Get a Price Jaw Crusher 100 150 Tph The production capacity of 150 TPH 200 TPH Stone Crusher Plant is popular in the world for the contractors and final users. flow chart of crusher plant ...Stone Crushing Plant,_0<< Get Price cost of coal crusher machine capacity of 10 .Crusher Stones In Canada, 2020-8-25 ·concrete crusher rental in canada, concrete crusher rental in canada2019/11/05 concrete crusher rental in canada coal crusher for sale samantha garfik shanghai jaw crusher what time indonesia ne ... 80-850T/H PROCESSED MATERIALS All kinds of mineral stone, except viscous materail. READ MORE. JC Series Jaw Crusher. Output size : 10mm to 275mm Production capacity : Up to 650TPH ...Jaw Crusher, Stone jaw crushing machine, Quarry Jaw ..., 2020-9-7 · Jaw Crusher is of high crushing ratio, larger capacity, well-distributed final product size, simple structure, reliable performance, convenient maintenance, lower operation cost, etc. It is widely used in mining, metallurgy, construction, highway, railroad, and water conservancy, etc.Step 2: PTH Crusher Series, 2020-9-2 ·.China PEX1030 Jaw Crusher (PEX, China PEX1030 Jaw Crusher (PEX-250x750), Find details about China Jaw Crusher, Stone Crusher from PEX1030 Jaw Crusher (PEX-250x750) - Shanghai Gator Machinery Co., Ltd.. Get Price PTH Multi Crusher, 2020-9-6 ·.Used Portable Stone Crusher Canada, 2020-9-9 · portablecone crusher manufacturer t h price, hammer crusher machines; how to run a cone crusher; stone cone crusher machine sand making stone quarry; industrial wood crusher; screener and crusher dealers in canada; used closed crusher plant in america; project of stone crusher in south africa; cs 5 1 2ft cone crusher dimensions; processing ore horizont al pump; stone crusher conveyor belt ...stone crusher bukaka kapasitas t/h, stone crusher bukaka kapasitas t/h d6d for sale nbsp018332dims l x w x h 3675 x 2390 x 2870mm weight 13150kgs updated tue july 14 2020 820 am emerald construction machinery equipment birmingham united kingdom b2 4ay seller information phone 44 121 368 1079 call phone call.Stone Crushing And Screening Line Crusher For Sale, stone crushing and screening line crusher for sale in. stone-crusher ...Cement Vertical Roller Mill Price Crusher For Sale, Used Vertical Cement Mill Small Rock Crusher On Sale. Cement grinding mill line in canada crusher for sale. Cement vertical roller mill price crusher for sale gtgt used mobile crusher in canada 4913 gtgt granite and basalt crushing line ton per day dry process cement grinding vertical roller mill for sale get price and support online cement grinding mill l. Get Price used stone crusher machine in ont canada, Used Used Stone Crusher Machine In Canada MINING Stone Crusher Machine In Canada Canada mobile stone crusher machine sale knm sales service stone crushers with a max working depth of 15 cm6 for tractors between 80 and 150 hp the stcl is - coal crusher 200tonnes per hour - manufacturing crusher plants in vietnam - dolomite crusher machine - spargo two roll ash crusher za in pretoria - 40 mt automatic aggregate crusher indi - digunakan portaable jaw crusher
https://jitkafrantova.it/industry/397/t-h-stone-crusher-in-canada.html
CC-MAIN-2021-25
refinedweb
599
56.89
Logging in the Script Task The use of logging in Integration Services packages lets you record detailed information about execution progress, results, and problems by recording predefined events or user-defined messages for later analysis. The Script task can use the Log method of the Dts object to log user-defined data. If logging is enabled, and the ScriptTaskLogEntry event is selected for logging on the Details tab of the Configure SSIS Logs dialog box, a single call to the Log method stores the event information in all the log providers configured for the task. For more information about logging, see Integration Services (SSIS) Logging. The following example demonstrates logging from the Script task by logging a value that represents the number of rows processed. using System; using System.Data; using Microsoft.SqlServer.Dts.Runtime; public class ScriptMain { public; } } } Blog entry, Logging custom events for Integration Services tasks, on dougbert.com
http://msdn.microsoft.com/en-us/library/ms136131.aspx
CC-MAIN-2014-35
refinedweb
151
53.21
So, Python is a high level, interpreted, general purpose programming language. It was created in 1991 by Guido Van Rossum. It emphasized on code readability. So, basically there are 2 types of language: high level and low level language. The low level language are close to Computer, so they are difficult to code but have other advantages. As python is a high level language so it is relatively easy to develop in python. Compied vs interpreted language There are 2 types of language on the basis of language on the basis of basically how we run the the program: compiled and interpreted languages. In the simple terms, compiled language are run in one go or the whole program will run in one time only if there are no errors. Below is a simple code in C++. note that i made 2 errors or mistakes on purpose. You don’t need it worry if you didn’t understand the code, basically there are 7 statements that prints a simple message. I made error in 5th and 6th line. #include <iostream> int main() { std:cout << ” hello world!”; cout< “welcome to betaPython”; std::cout << “by kashish Nagpal ” return 0; } Output: C++ (compiled language) output Note: that it shows 2 errors and didn’t produce any output. In compiled languages the program in fully analysed and compiler detect if there is any error or not .If there are any error then the compiler will give the error along wth line number. After removing all the errors the program will run or give output. Now below is a code in Python which is an interpreted language. basically it’s the same code that I have written in C++. it prints two statements but there are two errors in the program. a=6 print(a) print(b) print(betaPython by kashish Nagpal) Note: in this program, the interpreter first analyse the program line by line. it analyse first line then it give the output and analyse the second line and so on. So, it gives the output of the 1st line (no output) i.e. it prints the statements 6. As there is error in 3rd and 3th line but it only shows one error that is in the 3rd line. after we remove the error in the 3rd line, then it analyse the third line and if there is any error it will point it out. uses of Python - Web development - Data science - Video game development - Desktop GUI (graphical user interface) - Software development Pros /advantages of python - Easy to learn - Extensive libraries - Iot (internet of things) - Easy to develop software - One of the popular language - One of the widely used Programming language main products built using python - YouTube - Spotify Next article: how to install python in android, iOS, MAC OS, windows, Linux
https://betapython.com/introduction-to-python/
CC-MAIN-2022-05
refinedweb
467
71.24
Intro There are several posts online about automating your Keurig coffee maker, but few had all the features and details that I wanted. Therefore, I had borrowed some ideas from other Instructables and/or YouTube videos to come up with my own solution. What? The Keurig 30 will be equipped to turn on at a certain time to begin heating the water. It will reset the locking mechanism to accept the already-placed pod from the night before. Then, it will begin brewing an 8oz cup. Finally, it will turn off the machine. Why? Even the highest-priced Keurig models do not offer an auto-brew feature. While it makes sense, from an architecture standpoint, not to do this due to the need to change the pods every time you want to have a cup of coffee or tea - I still wanted to be able to wake up to my Keurig machine brewing. How? I decided to use Arduino-based architecture to tackle this project. - 1 - Arduino Uno (for prototyping the circuit board layout - not required) - 1 - Adafruit Feather Board () - Any microcontroller board should be fine, as long as it complies with the I2C protocol, and has 3 digital output pins - 3 - Adafruit non-latching relay () - Any similar relay is fine - 1 - Adafruit Real Time Clock module () - 1 or 2 - blank breadboards - 1 - computer running Arduino IDE - 1 – soldering iron, and solder wire - Misc – Safety glasses, well-ventilated work area, Dremmel, voltmeter, wires, electrical tape, hot glue gun, etc Step 1: Getting to the Control Board Getting to the control board Wow. There are lots of videos on YouTube on how this can be accomplished, and none claim this to be an easy task. I decided to forego these approaches, and instead resorted to using a Dremmel to cut the top lid off, thus revealing the control board. Don’t forget to unplug the Keurig from power before doing anything! As you can see depicted in the photos, the lid atop of the control pod is only being held in place by 3 posts. By cutting around these posts, you will be able to remove that portion of the lid. Then, a little glue and some electrical tape will allow you to cover up this mess once you are done. I was not aware of the exact locations of these posts, so your results will hopefully be much prettier than my experimental ones seen here. Alright, we have the control board in sight. It is held in place by a couple of screws. It is also connected to the control wires on the front, and the water pressure sensor tube on the bottom. I was able to gently pull the wires to allow for more slack, and was able to unplug the connector without the whole thing retracting. Slowly lifting the board will give you some space to gently disconnect the clear plastic tube of the pressure sensor (don’t worry, it will not retract into the depths of the machine either). Step 2: Extending the Control Board There we are - the control board! Simple enough. We only need access to the buttons for power and 8oz brew. The board is very well-labeled, so you should locate the two buttons and their corresponding +/- spots to extend with a few wires of your own. Once the wires are soldered on, you can hook everything back up and place the controller board back into its pod. You can test that everything is working properly by plugging the Keurig back in, and simply shorting the pair of wires associated with a function (i.e. Power). If everything is wired up properly, your Keurig should turn on without you having to push any buttons! Disclaimer – the last photo is not from my project, since I had closed it up before taking a picture, but this should give you an idea of where to connect your wires. Original source – Step 3: Extending the Lock Mechanism This part seems to be missing from most tutorials that I have seen. The premise here is that the Keurig will not allow you to place a pod into the machine then turn it off. After a short period of time, the Keurig will expect you to physically open the pod placement arm as if to reset itself and think that a new pod had been placed. We are going to override this via the Arduino. First, we need to remove the upper lid that sits on top of the brewing receptacle. There lies a mechanical button that gets pressed whenever the lid is closed. We are going to cut the two wires attached to the button, and remove the button altogether. Then we simply attach extension wires to the two liberated wires, and that is all for this step. The magic will happen in the code. Step 4: Setting Up Our Breadboard Hopefully, the pictures are self-explanatory. Basically, we wire up the RTC via 3v/GRD and SCL/SDA wires. We power up the relays via 3v/GRD. Then we connect each relay to the control board via any digital output pins. Each relay is, furthermore, connected to the wires extending the lock mechanism, the power button, and the 8oz button, respectively. The order doesn't matter as long as the code reflects which output pins you choose to use. Step 5: A Little Code – RTC Setup <p>// Date and time functions using a DS3231 RTC connected via I2C and Wire lib<br>#include <Wire.h> #include "RTClib.h"</p><p>RTC_DS3231 rtc;</p><p>char daysOfTheWeek[7][12] = {"Sunday", "Monday", "Tuesday", "Wednesday", "Thursday", "Friday", "Saturday"};</p><p>void setup () {</p><p>#ifndef ESP8266 while (!Serial); // for Leonardo/Micro/Zero #endif</p><p> Serial.begin(9600);</p><p> delay(3000); // wait for console opening</p><p> if (! rtc.begin()) { Serial.println("Couldn't find RTC"); while (1); }</p><p> if (rtc.lostPower()) { Serial.println("RTC lost power, lets set the time!"); rtc.adjust(DateTime(2017, 6, 11, 16, 38, 0)); // SET DATE/TIME HERE } }</p><p"); Serial.println(); delay(3000); }</p> Step 6: A Little (more) Code #include <Wire.h> #include "RTClib.h" RTC_DS3231 rtc; char daysOfTheWeek[7][12] = {"Sunday", "Monday", "Tuesday", "Wednesday", "Thursday", "Friday", "Saturday"}; bool power = false; int power_pin = 11; bool eight = false; int eight_pin = 9; bool lid = false; int lid_pin = 5; int alarm_h = 7; // Alarm hour int alarm_m = 00; // Alarm minute void setup () { Serial.begin(9600); pinMode(2, INPUT); pinMode(power_pin, OUTPUT); // ON/OFF pinMode(eight_pin, OUTPUT); // 8OZ CUP pinMode(lid_pin, OUTPUT); // LID delay(3000); // wait for console opening if (! rtc.begin()) { Serial.println("Couldn't find RTC"); while (1); } } void loop () { DateTime now = rtc.now(); Serial.print(now.hour(), DEC); Serial.print(':'); Serial.print(now.minute(), DEC); Serial.print(':'); Serial.print(now.second(), DEC); Serial.println(); // TURN KEURIG ON if (now.hour() == alarm_h && now.minute() == alarm_m){ if (!power){ Serial.print("HEATING..."); power = true; digitalWrite(power_pin, HIGH); delay(500); digitalWrite(power_pin, LOW); } } // RESET LID if (now.hour() == alarm_h && now.minute() == alarm_m + 3){ if (!lid){ Serial.print("RESETTING LID..."); lid = true; digitalWrite(lid_pin, HIGH); delay(1000); digitalWrite(lid_pin, LOW); digitalWrite(lid_pin, HIGH); delay(1000); // HIT 8OZ KEY if (!eight){ Serial.print("BREWING..."); eight = true; digitalWrite(eight_pin, HIGH); delay(500); digitalWrite(eight_pin, LOW); } } } //TURN KEURIG OFF if (now.hour() == alarm_h && now.minute() == alarm_m + 10){ if (power){ Serial.print("Done..."); digitalWrite(lid_pin, LOW); power = false; eight = false; lid = false; digitalWrite(power_pin, HIGH); delay(500); digitalWrite(power_pin, LOW); } } Serial.println(); delay(1000); } Step 7: Plug Everything Together That's about it! Make sure to change the code to run at a convenient time for testing. Then simply put in a coffee pod and enjoy your brew! Discussions 1 year ago I'd love to have an automatic system but I tend to drink too much coffee when I have my own coffee maker ^.^;
https://www.instructables.com/id/Keurig-Auto-brew/
CC-MAIN-2019-04
refinedweb
1,305
63.9
I've been doing quite a bit with user controls and JavaScript. There is an art to using the JavaScript correctly, wiring it in and having it work consistently. I thought I'd post a brief article on the "101" around this based on the things I've found. Hopefully this can be a quick-and-dirty reference guide for those of you working to extend your user controls with JavaScript functionality. Understand Multiplicity First, understand if your user control can potentially be rendered multiple times on a page. For the most part, I assume this will be the case. However, there are some circumstances, such as a large grid control, where you may absolutely know there is only going to be a single instance. The reason you want to understand this is because you want to try to have exactly one copy of the JavaScript code emitted to the browser, so it will need to be "aware" of the instance it is working with. JQuery I am a major fan of JQuery because of the fact it feels "natural" with the combination of XPath selector and CSS-style selections, chaining, and the browser compatibility is handled for me. There are correlating functions in other libraries, but this is what I'll be using for my examples. Part 1: The ASCX file I try to keep as clean a separation of code from the ASCX file as possible. The exception are elements that I must derive on the server side. These typically include references for callbacks, URLs (I like to strongly type my URL references), and wiring "init" events. We'll use a hypothetical "contact" control. Here is what I have at the bottom of my ASCX file, and this assumes I am using the AJAX framework and therefore have somehow registered a ScriptManager (this is outside the scope of this document — see the Script Manager control overview: <script type="text/javascript"> window.Sys.Application.add_init(contactInit); contactCtrlClientID = '<%=ClientID%>'; </script> Basically, I've registered with the AJAX framework and told it to call contactInit once everything is loaded. I've also created a reference to the client id. If I'm going to have multiples, I'd push them to an array or use some other method to keep track (windows provides handles and references in AJAX as well) of them ... this is just a quick and easy way to know how it is registered to the page if it is inside of a master page, etc. The JavaScript file There are many ways you can include the related JavaScript. Some people like to make it a straight include reference so they can easily update the JavaScript on disk and test changes, etc. We produce commericial software so the code is embedded (I like to do initial testing with an external resource, then embed the JavaScript). My favorite convention is to name the JavaScript the same as the control and have it side-by-side with the control in the designer (i.e. if my control is ContactCtrl.ascx, then my JavaScript is ContactCtrl.js and at the same level as the control). Embedding your JavaScript This is an easy three-step process. - Right click on the properties of your JavaScript file and indicate you wish to make it an Embedded Resource. If you don't want to ship the source, make sure you have "do not copy" selected for the destination. - Your web project should have a Properties folder with AssemblyInfo.cs. At the bottom of this file, you need to add a tag to identify the embedded resource. You will use the namespace, including the folders, and then the filename itself. For example, let's say the namespace for the project is Company.Project and the control is in a UserControls/Contact folder. Your resource will look like this: [assembly: WebResource("Company.Project.UserControls.Contact.ContactCtrl.js", "text/javascript")] - Finally, register a reference to it. When you register an embedded resource, you must specify a type. This ties the code to the control and prevents it from being emitted multiple times if the control appears multiple times. It is important to use typeof instead of GetType() or it may not behave as expected. Our example, which can go in onPreRender or onLoad: ... Page.ClientScript.RegisterClientScriptResource(typeof(ContactCtrl), ("Company.Project.UserControls.Contact.ContactCtrl.js"); ... Initialization When you register an initialize event, you can then hook in to "prep" the control. For example, maybe my contact control has some dynamic sections that are conditional based on a contact type. I can do something like this: function contactInit() { $(document).ready({ $("#divAdmin").hide(); $("#selType").change(contactTypeChange); }); } function contactTypeChange() { ... $("#divAdmin").show(); } Essentially, this will hide the admin section, then wire in a change event to the dropdown and call the type change method which can conditionally show it, etc. Events Good development practice dictates I design my control as self-contained as possible. For example, if I am making a search criteria control, I can't just assume I'll have a "search results grid" and when the user clicks "submit" send some client information over to the control. The proper way to have user controls talk on the client side is the same way on the server: through events. There are two ways I like to wire in my events. If I'm doing a light weight control, I'll tab into the global event pool and have a unique event name. If I'm developing a more extensive control, I'll wire in an AJAX behavior and expose events that way. To learn how to create a full-blown AJAX client control (something that can expose all of your server side events in a similar clientside model) read Creating Custom ASP.NET AJAX Client Controls. Let's assume I want to expose an event that is raised whenever the user changes the contact type. This way I can have other controls that might dynamically respond to the change on the client side, avoiding a server round trip. I don't have a true AJAX client control that corresponds to my contact control, so what can I do? An easy method is to plug into the Application events - think of these as "global" events. In my contact, I just need to know if someone is listening to my event. I'll give it the name "contactTypeChange" and send along the new type as well as my client id in case I have multiple instances. function contactTypeChange(clientID,contactType) { var contactHandler = window.Sys.Application.get_events().getHandler("contactTypeChanged"); if (contactHandler) { contactHandler(clientID,contactType); } } Now, in my control that is listening to the event, I simply register to do something with it. Let's call this my "Contact Security" panel and it will show special admin security rights if the contact is an admin. First, in my init function, I'll register for the event. Second, I'll provide a function to handle the event. function contactSecurityInit() { window.Sys.Application.get_events().addHandler("contactTypeChanged", contactTypeChangeHandler); } function contactTypeChangeHandler(clientID, contactType) { if (contactType == "admin") { $("#divSecurityAdminInfo").show(); } } Callbacks Finally, a caveat on callbacks. I use these a lot, because of the fact that update panels tend to add a lot of overhead. If your application requires javascript, you can manage a lot of the state in the browser page and use more tightly defined callbacks to perform dynamic updates. One "gotcha" with callbacks is that you must ask for the callback function in order for the control to listen for it. The three steps I find are easiest to manage callbacks: - Implement ICallbackHandler on your UserControl. This will generate two methods: one that takes an event argument, and one that asks for a return result. That's it! The event argument is what you control on the client side and pass down. In your control, you'll process some result and then can send back something to the client. It can be as simple as a little message to display, or as complex as a dynamic snippet of JavaScript to eval() on the client. - Ask for the callback reference. This is important! Even if you will explicitly callback from your code, you need to "ask" for the reference in order for the control to listen. It's as simple as: ... Page.ClientScript.GetCallbackEventReference(this, string.Empty, string.Empty, null); ... - Wire in the call! For example, let's say we want to display some text in a div based on the contact type. We want the server to retrieve the text and perform globalization functions against it before sending it back from the browser. This snippet of code will do the trick: WebForm_DoCallback(callbackRef, contactType, function(args, ctx) { $("#divHelp").html(args); }, null, null, false); Callbackref is the reference to the control. You can embed the GetCallbackEventReference above in your ASCX page and render the page to see how the control is referenced. It's essentially the path to the control separated by dollar signs (if you have a masterCtrl then a pageCtrl then the contactCtrl, it will look like masterCtrl$pageCtrl$contactCtrl). The next is the argument that gets passed to the server, following by the function to process when the result is returned. Here, we do an inline function and find the div and fill it with what we got back from the server. (You can read more about script callbacks). Conclusion Obviously, there is a lot you can do with scripts and user controls and this post only scratched the surface. Hopefully this gave you a good idea of some good practices for associating JavaScript to a user control and ways you can use JavaScript to extend the functionality, allow for greater interoperability between controls, and streamline performance through the use of callbacks.codeproject do you have a complete working source code to go with this article? Jeremy, I had to break a complicated page into user controls but I wasn't sure about the best way to decouple them yet still having them being able to communicate with each other. Your example of a search criteria control talking to a search results control through events was exactly what I was trying to do. This helped me very much. Thanks! Dave.
http://csharperimage.jeremylikness.com/2009/05/javascript-and-user-controls-101.html
CC-MAIN-2016-44
refinedweb
1,703
62.78
(For more resources related to this topic, see here.) Drawing a bar chart with Flex The Flex framework offers some charting components that are fairly easy to use. It is not ActionScript per say, but it still compiles to the SWF format. Because the resulting charts look good and are pretty customizable, we decided to cover it in one recipe. There is a downside though to using this: the Flex framework will be included in your SWF, which will increase its size. Future recipes will explain how to do the same thing using just ActionScript. Getting ready Open FlashDevelop and create a new Flex Project. How to do it... The following are the steps required to build a bar chart using the Flex framework. Copy and paste the following code in the Main.mxml file. When you run it, it will show you a bar chart. <.collections.ArrayCollection; [Bindable] private var monthsAmount:ArrayCollection = new ArrayCollection( [ { Month: "January", Amount: 35}, { Month: "February", Amount: 32 }, { Month: "March", Amount: 27 } ]); ]]> </fx:Script> <mx:BarChart <mx:verticalAxis> <mx:CategoryAxis </mx:verticalAxis> <mx:horizontalAxis> <mx:LinearAxis </mx:horizontalAxis> <mx:series> <mx:BarSeries </mx:series> </mx:BarChart> </s:Application> How it works... When you create a new Flex project, Flash Builder will generate for you the XML file and the Application tag. After that, in the script tag we created the data we will need to show in the chart. We do so by creating an ArrayCollection data structure, which is an array encapsulated to be used as DataProvider for multiple components of the Flex framework, in this case mx:BarChart. Once we have the data part done, we can start creating the chart. Everything is done in the BarChart tag. Inside that tag you can see we linked it with ArrayCollection, which we previously created using this code: dataProvider = "{monthsAmount}". Inside the BarChart tag we added the verticalAxis tag. This tag is used to associate values in the ArrayCollection to an axis. In this case we say that the values of the month will be displayed on the vertical axis. Next comes the horizontalAxis tag, we added it to tell the chart to use 10 as a minimum value for the horizontal axis. It's optional, but if you were to remove the tag it would use the smallest value in ArrayCollection as the minimum for the axis, so one month, in this case, March, would have no bar and the bar chart wouldn't look as good. Finally, the series tag will tell for a column, what data to use in ArrayCollection. You can basically think of the series as representing the bars in the chart. There's more... As we mentioned earlier, this component of the Flex framework is pretty customizable and you can use it to display multiple kinds of bar charts. Showing data tips Multiple options are available using this component; if you want to display the numbers that the bar represents in the chart while the user moves the mouse over the bar, simply add showDataTips = "true" inside the BarChart tag and it is done. Displaying vertical bars If you would like to use vertical bars instead of horizontal bars in the graph, Flex provides the ColumnChart charts to do so. In the previous code, change the BarChart tag to ColumnChart, and change BarSeries to ColumnSeries. Also, since the vertical axis and horizontal axis will be inverted, you will need verticalAxis by horizontalAxis and horizontalAxis by verticalAxis (switch them, but keep their internal tags) and in the ColumnSeries tag, xField should be Month and yField should be Amount. When you run that code it will show vertical bars. Adding more bars By adding more data in the ArrayCollection data structure and by adding another BarSeries tag, you can display multiple bars for each month. See the Adobe documentation at the following link to learn how to do it:. Building vertical bar charts Now that we have built a bar chart using Flex, we are ready to do the same in pure ActionScript. This bar chart version will allow you to expand it in multiple ways and will remove the weight that the Flex framework adds to the file size. Now a bit about bar charts; Bar charts are good when you don't have too much data (more than 20 bars starts to make a big chart), or when you've averaged it. It is a quick way to compare data visually. Getting ready All we will need for this is to start a new project in FlashDevelop. Also, it would help to read about preparing data and about axes in the book ActionScript Graphing Cookbook. How to do it... This section will refer a lot to the code provided with the book. You will notice that we divided all the elements in the charts into their own classes. It all starts in the Main.as file, where we create the data that we will use to display in the chart after that we just create the chart and add it to the display list. var data:Vector.<BarData> = new Vector.<BarData>(); data.push(new BarData("January", 60)); data.push(new BarData("February", 100)); data.push(new BarData("March", 30)); var chart:BarChart = new BarChart(data, 400, 410); chart.x = 30; chart.y = 30; addChild(chart); From here you can look into the BarData class, which it is just two variables, a string and a number that represents the data that we are going to show. We now need to create a class for all the elements that comprise a bar chart. They are: the bars, the vertical axis, and the horizontal axis. Now this recipe is building a vertical bar chart so the vertical axis is the one that will have numerical marks and the horizontal axis will have labels on the marks. First the Bar class: This class will only draw a rectangle with the height representing the data for a certain label. The following is its constructor: public function Bar(width:int, height:int) { graphics.beginFill(0xfca25a); graphics.drawRect(-width/2, 0, width, -height); graphics.endFill(); } The horizontal axis will take the x coordinate of the created bars and will place a label under it. public function HorizontalAxis(listOfMark:Vector.<Number>, data:Vector.<BarData>, width:Number) { drawAxisLine(new Point(0, 0), new Point(width, 0)); for (var i:int = 0; i < listOfMark.length; i++) { drawAxisLine(new Point(listOfMark[i], -3), new Point(listOfMark[i], 3)); var textField:TextField = new TextField(); textField.text = data[i].label; textField.width = textField.textWidth + 5; textField.height = textField.textHeight + 3; textField.x = listOfMark[i] - textField.width / 2; textField.y = 5; addChild(textField); } } Now the vertical axis will make 10 marks at regular interval and will add a label with the associated value in it: for (var i:int = 0; i < _numberOfMarks; i++) { drawAxisLine(new Point( -3, (i + 1) * -heightOfAxis / _ numberOfMarks ), new Point(3, (i + 1) * -heightOfAxis / _ numberOfMarks)); var textField:TextField = new TextField(); textField.text = String(((i + 1) / (_numberOfMarks)) * maximumValue ); textField.width = textField.textWidth + 5; textField.height = textField.textHeight + 3; textField.x = -textField.width - 3; textField.y = (i + 1) * -heightOfAxis / _numberOfMarks - textField.height / 2; addChild(textField); } Finally, the BarChart class will take the three classes we just created and put it all together. By iterating through all the data, it will find the maximum value, so that we know what range of values to put on the vertical axis. var i:int; var maximumValue:Number = data[0].data; for (i = 1; i < data.length; i++) { if (data[i].data > maximumValue) { maximumValue = data[i].data; } } After that we create each bar, notice that we also keep the position of each bar to give it to the horizontal axis thereafter: var listOfMarks:Vector.<Number> = new Vector.<Number>(); var bar:Bar; for (i = 0; i < data.length; i++) { bar = new Bar(_barWidth, data[i].data * scaleHeight); bar.x = MARGIN + _barSpacing + _barWidth / 2 + i * (_barWidth + _barSpacing); listOfMarks.push(bar.x - MARGIN); bar.y = height - MARGIN; addChild(bar); } Now all we have left to do is create the axes and then we are done; this is done really easily as shown in the following code: _horizontalAxis = new HorizontalAxis(listOfMarks, data, width - MARGIN); _horizontalAxis.x = MARGIN; _horizontalAxis.y = height - MARGIN; addChild(_horizontalAxis); _verticalAxis = new VerticalAxis(height - MARGIN, maximumValue); _verticalAxis.x = MARGIN; _verticalAxis.y = height -MARGIN; addChild(_verticalAxis); How it works... So we divided all the elements into their own classes because this will permit us to extend and modify them more easily in the future. So let's begin where it all starts, the data. Well, our BarChart class accepts a vector of BarData as an argument. We did this so that you could add as many bars as you want and the chart would still work. Be aware that if you add many bars, you might have to give more width to the chart so that it can accommodate them. You can see in the code, that the width of the bar of determined by the width of the graph divided by the number bars. We decided that 85 percent of that value would be given to the bars and 15 percent would be given to the space between the bars. Those values are arbitrary and you can play with them to give different styles to the chart. Also the other important step is to determine what our data range is. We do so by finding what the maximum value is. For simplicity, we assume that the values will start at 0, but the validity of a chart is always relative to the data, so if there are negative values it wouldn't work, but you could always fix this. So when we found our maximum value, we can decide for a scale for the rest of the values. You can use the following formula for it: var scaleHeight:Number = (height - 10) / maximumValue; Here, height is the height of the chart and 10 is just a margin we leave to the graph to place the labels. After that, if we multiply that scale by the value of the data, it will give us the height of each bar and there you have it, a completed bar chart. There's more... We created a very simple version of a bar chart but there are numerous things we could do to improve it. Styling, interactivity, and the possibility of accommodating a wider range of data are just some examples. Styling This basic chart could use a little bit of styling. By modifying the color of the bars, the font of the labels, and by adding a drop shadow to the bars, it could be greatly enhanced. You could also make all of them dynamic so that you could specify them when you create a new chart. Interactivity It would be really good to show the values for the bars when you move the mouse over them. Right now you can kind of get an idea of which one is the biggest bar but that is all. If this feature is implemented, you can get the exact value. Accommodating a wider data range As we explained earlier, we didn't account for all the data range. Values could be very different; some could be negative, some could be very small (between 0 and 1), or you would want to set the minimum and maximum value of the vertical axes. The good thing here is that you can modify the code to better fit your data. Creating comparison bar charts In the previous recipe, we learned how to make a simple bar chart. The relation of the data was one-to-one; one label, one value. Now we will make this a bit more complex by adding one more dimension. We will create a bar chart in which the bars are replaced by bar charts. It can be hard to understand it like this, but it gets clearer if you look at the data. For this recipe we will compare the medals (bronze, silver, and gold) won by three different countries at a sporting event. Getting ready A good starting point is to copy over the project from the previous recipe (see the Building vertical bar charts recipe). We will modify it a lot, but the core of it will be similar. How to do it... The following are the steps required to build a comparison bar chart: As usual, when we create a new chart, we always start with the data. So we create a new class for it called ComparisonChartData. Check the file out; it is basically a label and a vector of BarData from the previous project. After that, in the main class we will generate the data. This example will use the number of medals won for countries as data. After that we create another chart. Now if you look at the classes VerticalAxis.as, HorizontalAxis.as, and Bar.as, you will notice that they have been only slightly modified. The HorizontalAxis class has been made a bit more data agnostic and requires a list of labels instead of being passed the data, that way we will be able to use the same class for bar chart and the comparison bar chart. The Bar class is now passed a color so that we can specify the color of the bar. Most of the changes happen in the ComparisonBarChart class. First we must determine the size of the bars: _categoryWidth = (chartWidth - MARGIN) * (1 - CATEGORY_PERCENT) / data.length; _categorySpacing = (chartWidth - MARGIN) * CATEGORY_PERCENT / (data.length + 1); _barWidth = _categoryWidth * (1 - BAR_PERCENT) / data[0].bars. length; _barSpacing = _categoryWidth * BAR_PERCENT / (data[0].bars.length – 1); After that we loop over the data and create all the bars: for (i = 0; i < data.length; i++) { var markX:Number = _categorySpacing + _categoryWidth / 2 + i * (_categoryWidth + _categorySpacing); listOfMarks.push(markX); listOfLabels.push (data[i].label); for (j = 0; j < data[i].bars.length; j++) { bar = new Bar(_barWidth, data[i].bars[j].data * scaleHeight, data[i].bars[j].color); bar.x = MARGIN + markX - _categoryWidth / 2 + j * (_barWidth + _barSpacing) + _barWidth / 2; bar.y = chartHeigth - MARGIN; addChild(bar); } } How it works... As we mentioned in the Introduction, we added a dimension to the chart so the data must reflect it. In this case we reuse the BarData class we had in the previous class, which is still very relevant (our chart will still have bars) but we create another class (ComparisonChartData) to hold a list of BarData. After that the only hard part is determining the space allowed for each category (in this example: each country). We use the same formula as for determining the size of the bars in the BarChart recipe. Each category will be subdivided into multiple bars. So from the category width we can now fnd the width of the bars. We also took the liberty of assigning 15 percent for the space between the bars. Finally we loop over all the data. Since our data is more complex now, we must do a loop inside of another loop to create all the bars. We find the position of the marks inside the first loop much like we found the position of the bars in the BarChart class. And from that position, we find the position of each of the bars for each of the categories. There's more... We added complexity to the BarChart code and created the ComparisonBarChart. In the following section we will look at how to improve it. Adding a legend Not having a legend for the previous chart was fine because you could have understood what it was depicting just by looking at it. But in the case of more complex data, a legend could really help explaining what we are trying to show with the chart. Fortunately, creating a legend is really easy. You should create a class for it, which you could re-use in other charts or graph. Drawing histograms Well truth be told, we won't really be making histograms with this recipe but mostly bar charts in the style of histograms. Histograms are very rigid in their nature and follow really specific mathematical functions to represent frequency and density. You can read more on histograms on Wikipedia:. What we are going to do here is build a histogram styled bar chart (no space between the bars) that uses a function as input. You can think of this as a different way to do an area chart or a line chart. Getting ready The data for this recipe is a bit different as it uses a mathematical function instead of a data set. You can get a mathematical function using interpolation or by having a program such as Microsoft Excel do it for you. How to do it... The following are the steps required to build a histogram using ActionScript 3: Now if you look at our Main.as class, you will notice that it is quite small compared to the previous recipe. For this recipe, we don't need to create data; we will use a function to get our data. The function used in this case is the one as follows: public function dataFormula(xValue:Number):Number { return (xValue * xValue * xValue); } Once we have this function we can create the chart: var chart:Histogram = new Histogram(dataFormula, 400, 410, 25); Next, in the Bar.as file we will add a style to the line because the bars will be really close in this chart and this will help differentiate them. Finally, the Histogram.as file is different than BarChart.as and takes a function as a parameter in the constructor. From that function, we will determine the maximum value and we will create the bars. var maximumValue:Number = data(0); for (i = 1; i < numberOfBars; i++) { if (data(i) > maximumValue) { maximumValue = data(i); } } var scaleHeight:Number = (chartHeight - 10) / maximumValue; var listOfMarks:Vector.<Number> = new Vector.<Number>(); var listOfLabels:Vector.<String> = new Vector.<String>(); var bar:Bar; for (i = 0; i < numberOfBars; i++) { bar = new Bar(_barWidth, data(i) * scaleHeight); bar.x = 10 + _barWidth / 2 + i * (_barWidth); listOfMarks.push(bar.x - 10); listOfLabels.push(String(i)); bar.y = chartHeight - 10; addChild(bar); } How it works... Contrary to the other recipe, this one won't use a list of numbers as data but a function in the code called dataFunction. This example will use the function y = x3. The value on the vertical axis is equal to the value on the horizontal axis multiplied by itself two times (to the power of three). But you could use any function that takes a number and returns one. Since we want the graph to be in the style of a true histogram, we don't need to compute any space between the bars. To get the width of the bars, we take the width of the graph divided by the number of bars plus one. We add one to the number of bars so that the last bar doesn't come too close to the end of the graph on the right-hand side. We are now ready to use our function to get our data. Since we know it takes a number and returns a number, by giving it our value of the horizontal axis it will return the value of the vertical axis. The only code we have to write to do so is data (horizontalAxisValue). In our case that value is the index of our loop. Note that we could have modifed the HorizontalAxis class to be like the VerticalAxis class, but it still worked (the labels weren't too big) so we left it as it is. There's more... Accepting a wider data range would improve the Histogram class, but also, you could use it in a different way by making it draw a lot of bars. Data range Here again the data range is very important. It would be good to modify the code for this recipe so that it could accept a minimum value and a step value. In this example, our minimum was 0 and our step value was 1. But that may not fit for every function or data. So many bars One interesting thing to note here is that if you set the number of bars to be very high, you will end up creating an area chart since the bars will be really thin. If you want to use this chart in this way, I would suggest modifying the HorizontalAxis class because we used it as it is from the BarChart recipe and it would create way too many labels. Creating sparklines to enrich text content Sparklines are small line charts that are used to augment the value of some other content by giving a visual trend of a specific metric. One of the easy examples of applications that use sparklines is Google Analytics. As you can see in the previous screenshot, the sparklines on the left give you a clearer background about the numbers on the right. So as of today, you had 13,935 visits but those varied in a specifc pattern in the previous month. Getting ready Just start a new project in FlashDevelop. This will be a really simple recipe. How to do it... The following are the steps required to build sparklines: Since sparklines are a pretty simple graph, we only have two classes for this recipe. In the main class, we generate the data and instantiate the Sparkline class. All of the drawing is done inside Sparkline.as. We will only use the graphics function of the Sprite class to draw the line and the axes: graphics.lineStyle(1, 0x0000ff); graphics.moveTo(0, chartHeight-data[0] * scaleHeight); for (i = 1; i < data.length; i++) { graphics.lineTo(i * _stepSize, chartHeight-data[i] * scaleHeight); } graphics.lineStyle (1, 0); graphics.moveTo(0, chartHeight); graphics.lineTo(0, 0); graphics.moveTo(0, chartHeight); graphics.lineTo(chartWidth, chartHeight); How it works... Using the Flash Drawing API, the previous code will draw the two axes and draw a line that will represent our data. We based the data on visits to a website for a day. We randomly assigned values ranging from 9000 to 15500. This graph will give us an idea of how the traffic varied during the previous month. The thing to notice here is that since this line graph is so small, there are no labels anywhere. The goal here is not to check a precise data point, but more to derive a trend from the graph. We still need, like in the other graph, to determine the highest value in the data set to convert values to pixels. In this case, since we didn't provide x and y coordinates, we use the index of the vector as the horizontal coordinate; that is why we need to find the stepSize value. After that, we can just iterate over the data and draw a succession by using the lineTo calls. for (i = 1; i < data.length; i++) { graphics.lineTo(i * _stepSize, chartHeight-data[i] * scaleHeight); } From there the job is mostly done; we only need to draw the axes. This is done easily by the lineTo calls since we don't need to add labels. There's more... By coloring the area under the line and averaging the data we could improve on the sparkline recipe. Adding the area under the line The graphs from Google Analytics also have the area under the line drawn, so if we wanted to replicate that it wouldn't be too hard. We would have to make a second for loop before the one we already have and add a beginFill call before it. After that we have to make sure we close the shape so that the fill is complete. We draw the fill before so that is doesn't hide the line. You could also do it all in one loop if you started lineStyle just before drawing the top side of the area and if you remove it just after. Averaging the data As you can see, the line is a bit jerky in the graph. Since we are more interested in the trend with sparklines, you could average the data for every two or three values. This would give you fewer data points and you would lose some precision, but it would, in return, give you a smoother line and it might be easier to see the trends. Making 3D bar charts In the recipe Building vertical bar charts in this article, we showed how to make a simple bar chart in ActionScript. For this recipe, we wanted to spice it up a bit and make it 3D. How to do it... The following steps will show you how to convert a 2D bar chart to 3D: Take BarData.as, HorizontalAxis.as, VerticalAxis.as, and Main.as from the Building vertical bar charts recipe from the code files provide with the book; these files are not going to change. Now in BarChart.as, increase the space between the bars. These are the lines that change: _barWidth = (width - 10) * 70 / 100 / data.length; _barSpacing = (width - 10) * 30 / 100 / (data.length + 1); The big changes will be in the Bar.as file. First we will add two helper functions to go back and forth between the color model, as shown in the following code snippet: private function _hexToRGB(hex:uint):Object { var rgbObj:Object = { red: ((hex & 0xFF0000) >> 16), green: ((hex & 0x00FF00) >> 8), blue: ((hex & 0x0000FF)) }; return rgbObj; } private function _RGBtoHEX(red:int, green:int, blue:int):uint { return red << 16 | green << 8 | blue; } We will need to compute a new color for the top and right part. The following code will show us how we get the color for the top part: var baseColor:Object = _hexToRGB(color); baseColor.red += baseColor.red * 0.1; if (baseColor.red > 255) { baseColor.red = 255; } baseColor.green += baseColor.green * 0.1; if (baseColor.green > 255) { baseColor.green = 255; } baseColor.blue += baseColor.blue * 0.1; if (baseColor.blue > 255) { baseColor.blue = 255; } var topColor:uint = _RGBtoHEX(baseColor.red, baseColor.green, baseColor.blue); All that is left to do, is draw two parallelograms that will make our bar look 3D: graphics.beginFill(topColor); graphics.moveTo(-width / 2, -height); graphics.lineTo(-width / 2 + xOffset, -height - yOffset); graphics.lineTo(width / 2 + xOffset, -height - yOffset); graphics.lineTo(width / 2, - height); graphics.lineTo( -width / 2, -height); graphics.beginFill(rightColor); graphics.moveTo(width / 2, 0); graphics.lineTo(width / 2 + xOffset, 0 - yOffset); graphics.lineTo(width / 2 + xOffset, -height - yOffset); graphics.lineTo(width / 2, - height); graphics.lineTo(width / 2, 0); How it works... We are basically taking every class from the Building vertical bar charts recipe, unchanged, except for Bar.as and BarChart.as. In BarChart.as, we made a small modification to leave more space between our bars. Indeed, we will need space to have our 3D representing shapes. Before, we had 15 percent of space between bars; we will up that to 30 percent (in step 2). The meaty part is in Bar.as. We will add two parallelograms to simulate the 3D effect, one for the top and the other for the side to give a sense of depth. Those are pretty simple to draw; we find the x and y of the off-setted top part of our parallelogram (once you have this point you can use it to find all the other points of the shape) and using the lineTo function, we can draw lines until we are back to the starting position (that is, step 6). That already gives us a 3D looking shape, but there is a small problem. Since every piece is the same color, it reduces the 3D effect. That is why we are going to change the color of the top and right parallelograms. The top one is going to be of a lighter color and the right part is going to be of a darker color than the color specified in the parameters. To compute those colors, we will need to decompose the hex colors provided into red, green, and blue. For each of those components, we are going to add or remove 10 percent of their value (add for lighter, remove for darker) thus keeping a similar color. That is exactly what we do in step 5. Finally we convert it back to hexadecimal, as it is the format that the beginFill function expects. Now our 3D bar chart looks great! There's more... By using gradients and a real 3D engine you could improve on this recipe. Using gradient Another option to using different color tones and to accentuate the 3D effect, would be to use a gradient on the top and right parts. It would make your bar chart look a bit more refined. Using a real 3D engine Here we faked the 3D, but to get really good results that are easier to animate or to interact with, you should use a real 3D engine such as Away3D, which you can find here:. Summary This article gave us all the tools to create multiple bar charts, be it vertical, horizontal, comparison, or histograms, it provided us with solid bases to improve those charts even more. Resources for Article : Further resources on this subject: - Dynamic Flash Charts - FusionCharts style [Article] - Customizing Graphics and Creating a Bar Chart and Scatterplot in R [Article] - Graphical Capabilities of R [Article]
https://www.packtpub.com/books/content/creating-bar-charts
CC-MAIN-2015-18
refinedweb
4,901
64.61
Table of contents Created 16 March 2009 Requirements Prerequisite knowledge This article assumes intermediate to advanced knowledge of Flex, ActionScript 3.0, and MXML. You should know how to create Adobe AIR and Flex projects in Flex Builder 3, and be comfortable with a few object-oriented concepts such as inheritance and interfaces, as well as a splash of UML 2.0. User level Intermediate Adobe Flash Player and Adobe AIR provide the ability to run full-featured rich Internet applications (RIAs) either inside a web browser or as native desktop applications on both Windows and Mac OS X, with desktop Linux support coming sometime after the release of Adobe AIR 1.0. The problem for those who want to develop applications that support both desktop and web runtime environments is that Adobe AIR includes extra APIs that allow for local file access, SQLite database support, native windows, and more. These APIs won't compile into a traditional Flex web-based application destined for web-browser distribution (SWF files that you place on a web server and run inside the browser). In this article, I'm going to show you a technique to set up your Flex Builder 3 workspace and organize your code to output both a Flex web and desktop Adobe AIR application from the same code base. This technique enables you to create a single maintainable code base that will give you web and Adobe AIR applications from your Flex Builder workspace. To demonstrate this, I will set up three projects: one for Adobe AIR, one for the web, and one for the common code. The web and desktop applications will then make calls against an interface that is defined in a common project, but has concrete implementations in each of the desktop and web projects, allowing the desktop version to use the new functionality of Adobe AIR, and giving the developer the option to support a different implementation for the web-based application. With this knowledge, you'll be able to write software for both a rich browser experience and a desktop application from a code base that doesn't have a bunch of duplicate code. In doing this, you could offer both an online demo with crippled functionality and a desktop version of the application that can run offline with more advanced functionality. As a developer, if you tried to compile a Flex application destined for web-browser distribution that referenced shared code with dependencies on the Adobe AIR libraries, you'd be out of luck with compile-time errors. To get around this, you need to have a specific project for your Flex web output, and another for your Adobe AIR output. On top of this, because you don't want to duplicate all your code in both projects, you need a third project that contains all the common code of the application. Then your common code makes calls to an interface that has abstracted out the functionality that relies on specific runtime functionality (desktop versus web). Don't worry, this sounds a lot worse than it is. The easiest way to understand where to start is with the big picture of the solution. Look at Figure 1, which is a UML diagram of the sample project. I'll get down to the nitty-gritty of the code in the following sections. It's ok if you don't fully understand everything at this point; it's natural for this to take some time to sink in. Hopefully by the time you're done looking at the code, you'll be a guru of this implementation. The core application code lives in the Common Project. Adobe AIR specific and Flex specific code live in their own projects. The main application is MainCanvas.mxml and this needs to be added to both the Adobe AIR Project and Web Project with a call to addChild, just like adding any regular Flex control to a container from within ActionScript. Don't think of your WindowedApplication (for Adobe AIR) or Application ( for Flex Web Project) as your entry point to start coding. These are only shells that host the application. Next, to call a method that's been extracted, like saveFile, the CommonCode must call against the IGeneral interface that defines the functionality in mind. To get at an implemented version of saveFile, the CommonCode must get a reference to IGeneral from the GeneralFactory object, which retuns objects that implement IGeneral, supporting the saveFile functionality. GeneralFactory is self-aware of whether it's running inside Adobe AIR or a regular Flex application and will return either FlexGeneralImplementation or AirGeneralImplemenation, depending on the environment it's running in. When compiling the Adobe AIR desktop version of the application, AirGeneralImplementation is compiled into the final application with all its links to the Adobe AIR specific APIs. When compiling the Flex version into a SWF for running inside a web browser, FlexGeneralImplementation (without any references to the Adobe AIR APIs) is compiled into the application. The GeneralFactory returns the appropriate class to the CommonCode application where the appropriate saveFile method is called. The Adobe AIR Project will use browseForSave, an API specific to Adobe AIR, to save a file locally. The Web Project's implementation just flashes a quick message informing you to download the full version of the application. Here's how calling abstracted functionality looks in the main application code: var general:IGeneral; general = GeneralFactory.getGeneralInstance() general.saveFile(); That's it; pretty straightforward after you've organized all the code and created the appropriate classes. Figure 1, on the previous page, shows the code elements you will create in your application to implement the following pattern: - The CommonProject project contains elements used by both the AIR and the web project. It contains the MainCanvas.mxml user interface component. This project abstracts any functionality that relies on Adobe AIR specific API support into an interface ( IGeneral). This includes functionality such as saving files or writing to SQLite. The GeneralFactory class is a Factory that knows how to create the Adobe AIR and web-specific concrete classes. - The AirProjectproject contains elements specific to the AIR project. It also references the MainCanvas.mxml component, defined in CommonProject. - The. Setting up the Flex Builder 3 workspace and projects To get started, you'll need to create three projects in a new Flex Builder workspace. I'm assuming you're already familiar with creating projects in Flex Builder. (If not, visit the Flex Developer Center to get started with Flex.) - Create a new Flex Builder workspace. -. - From the ZIP source files provided with this article, copy files in the the CommonProject subdirectory into the directory for your CommonProject Flex project. - Create a Flex project named AirProject. Create this as a Desktop (AIR) application. During project creation, you'll be prompted with the ability to add a source path. Click Add Folder and browse to the src subfolder in the folder containing the CommonProject(which you created in step 2). This will share the source code from the common project (which will be important later on). Figure 2 shows what an appropriately added source path looks like. This project will have full access to Adobe AIR APIs. - From the ZIP source files provided with this article, copy files in the the AirProject subdirectory into the directory for your AirProject Flex project. - Create a third project, called WebProject; this should be a regular Flex project that compiles to a simple SWF file for web distribution. Just like the second project above, make sure you add the src subdirectory of the CommonProject project as a Source path (see Figure 2). This project will not have access to the Adobe AIR APIs. - From the ZIP source files provided with this article, copy files in the the WebProject subdirectory into the directory for your WebProject Flex project. Adding the Common UI component. Creating the IGeneral interface. Using the factory patterninterface, where the functionality is defined to provide either Adobe AIR or web-specific implementations: import IGeneral; import flash.system.ApplicationDomain; import flash.system.Security; - Define the GeneralFactoryclass: public class GeneralFactory { public static const NOT_SUPPORTED_IN_WEB_MESSAGE:String = "Not supported in web version."; - Create the getClassToCreatewhich dynamically inspects the current ApplicationDomain. This grabs the definition of the class to create, which is either the AirGeneralImplementationis only defined in the Adobe AIR framework, use the string "application"instead. static public function get isAir():Boolean { return Security.sandboxType.toString() == "application" ? true : false; } Implementing the Adobe AIR and Flex Concrete classes"); } } Calling the implemented); } } Where to go from here: - Decoupled functionality that provides different functionality for web and desktop versions or your application. A desktop version saves the file locally; the online version publishes a file to an online photo storage service, such as Flickr, for example. Or, the desktop version writes data to the local SQLite database and syncs up with online storage when connected; the online version saves directly to online storage. - Selective functionality in online version, offering an upgrade path and incentive for users to purchase a full desktop version of the application. This could be useful from a business standpoint (improving cash flow) or from a technical standpoint (like using data from another website that doesn't support cross-domain policy files). - Simplified development; implement complex functionality in your Adobe AIR application first, eventually rolling it into your Flex version a few releases later. - Minimal code duplication, because you have only one code base. - Option to offer an online demo version of the software that users can try without installing any software. Then provide an upgrade path to full-featured desktop application available for purchase. This is value-added proposition that only Adobe supports. No one else can simply offer this in their development platform. This is great for users (no application to install), and great for the company with potential users trying your application out in a matter of seconds..
https://www.adobe.com/devnet/air/flex/articles/flex_air_codebase.html
CC-MAIN-2018-39
refinedweb
1,650
52.29
Many multiplayer games can use the Network Manager to manage connections, but you can also use the lower-level NetworkServer and NetworkClient classes directly. When using the High-Level API, every game must have a host server to connect to.. That player’s instance of the game runs a “local client” instead of a normal remote client. The local client uses the same Unity Scenes and GameObjects as the server, and communicates internally using message queues instead of sending messages across the network. To HLAPI code and systems, the local client is just another client, so almost all user code is the same, whether a client is local or remote. This makes it easy to make a game that works in both multiplayer and standalone mode with the same code. A common pattern for multiplayer games is to have a GameObject that manages the network state of the game. Below is the start of a NetworkManager script. This script would be attached to a GameObject that is in the start-up Scene of the game. It has a simple UI and keyboard handling functions that allow the game to be started in different network modes. Before you release your game you should create a more visually appealing menu, with options such as “Start single player game” and “Start multiplayer game”. using UnityEngine; using UnityEngine.Networking; public class MyNetworkManager : MonoBehaviour { public bool isAtStartup = true; NetworkClient myClient; void Update () { if (isAtStartup) { if (Input.GetKeyDown(KeyCode.S)) { SetupServer(); } if (Input.GetKeyDown(KeyCode.C)) { SetupClient(); } if (Input.GetKeyDown(KeyCode.B)) { SetupServer(); SetupLocalClient(); } } } void OnGUI() { if (isAtStartup) { GUI.Label(new Rect(2, 10, 150, 100), "Press S for server"); GUI.Label(new Rect(2, 30, 150, 100), "Press B for both"); GUI.Label(new Rect(2, 50, 150, 100), "Press C for client"); } } } This basic code calls setup functions to get things going. Below are the simple setup functions for each of the scenarios. These functions create a server, or the right kind of client for each scenario. Note that the remote client assumes the server is on the same machine (127.0.0.1). For a finished game this would be an internet address, or something supplied by the Matchmaking system. // Create a server and listen on a port public void SetupServer() { NetworkServer.Listen(4444); isAtStartup = false; } // Create a client and connect to the server port public void SetupClient() { myClient = new NetworkClient(); myClient.RegisterHandler(MsgType.Connect, OnConnected); myClient.Connect("127.0.0.1", 4444); isAtStartup = false; } // Create a local client and connect to the local server public void SetupLocalClient() { myClient = ClientScene.ConnectLocalServer(); myClient.RegisterHandler(MsgType.Connect, OnConnected); isAtStartup = false; } The clients in this code register a callback function for the connection event MsgType.Connect. This is a built-in message of the HLAPI that the script invokes when a client connects to a server. In this case, the code for the handler on the client is: // client function public void OnConnected(NetworkMessage netMsg) { Debug.Log("Connected to server"); } This is enough to get a multiplayer application up and running. With this script you can then send network messages using NetworkClient.Send and NetworkServer.SendToAll. Note that sending messages is a low level way of interacting with the system. Did you find this page useful? Please give it a rating:
https://docs.unity3d.com/2017.4/Documentation/Manual/UNetClientServer.html
CC-MAIN-2020-05
refinedweb
545
57.47
hi, I want to intrupt a for loop which count till 30 or 60. I dont want to break the loop. If any intrupt(use of any key) occurs it should come out of the loop.. can anyone suggest me. Post your Comment using switch,break and for loop using switch,break and for loop generate a 10 digit number and display the length of longest increasing series C Break for loop C Break for loop In this section, you will learn how to use break statement in a for loop. The break statement terminates the execution of the enclosing loop Java break for loop Java break for loop  ... baadshah. In the example, a labeled break statement is used to terminate for loop... control outside the outermost loop. Break loop to output the following code using a loop the code is as follows with the user...(); if(code.equals("XX")){ break...(); if(code.equals("XX")){ break; } System.out.print("Quantity While loop break causing unwanted output While loop break causing unwanted output Below I will paste my code. My problem is that when I run this I get to input a int and if correct...{ System.out.println("You are correct");} break PHP Break Break Control Structure: Break is a special kind of control structure which helps us to break any loop or sequence, like if we want to break the flow of any loop (for, while etc) then we can use break statement, generally we need break use break. It gives the following output: C:\javac>... in the following example we have used break statement. In this the inner for loop... break   Java Break out of for loop Java Break out of for loop  ... javax.swing.JOptionPane; public class Java_Break_out_of_for_loop..., "Tip of the day.. \n\tAllow the label 'break' to close 'for' loop", "Welcome Java Break loop Java Break loop  ... statement is used to break two looping statements do-while & for loop... in terminating the loops. Break Loop Example Java Break to the next statement following the loop statement. Break is one of the controlling statement like continue and return. Break statement can be used in while loop... Java Break Many Continue and break statement loop whereas the break statement causes the control outside the loop. Here... then break statement will call and control will jump outside the current loop... loop whereas the break statement causes the control outside the loop. Here C break continue example C break continue example  ... to force an immediate jump to the loop control statement. The break statement... negative value and the break statement terminates the loop. Download Source Code Java Break while example ; Break is often used in terminating for loop but it can also... label mass. Break While Loop Java... break the \n\t\twhile loop \n\t\tSuccessfully", "Congrats ", 1);  Java Break Statement Java Break Statement 'break' statement comes under the branching statements category of java. It is used to terminate the loop coded inside the program. Example Break statement in java happen that we want to come out of the loop in that case break is used...; loop. Within a loop if a break statement is encountered then control resume...) break; } } } In the above example inside a for loop Break Statement in JSP are using the switch statement under which we will use break statement. The for loop...Break Statement in JSP The use of break statement is to escape The break Keyword . Also break keyword is used for terminate a loop. The break always exits... The break Keyword  .... In other word we can say that break keyword is used to prematurely exit Java Break continue category. 'break' allows users to give end to a loop whereas with 'continue...]); i++; if(c[i] == c_1) break; else continue... Java Break continue   Java Break Lable Java Break Lable In Java, break statement is used in two ways as labeled and unlabeled. break is commonly used as unlabeled. Labeled break statement is used Java - Break statement in java break statement and exit from the loop and loop is terminated. The break...;javac Break.java C:\chandan>java Break The Prime number in between 1 - 50... Java - Break statement in java   Break Statement in java 7 terminate the loop like while, do-while, for and switch. When you call break... break statement.."); break; // using break statement to break for loop...Break Statement in java 7 In this tutorial we will discuss about break What is the difference between a break statement and a continue statement? to exit from a loop before the completion of the loop program we uses the break...What is the difference between a break statement and a continue statement? Hi, What is the difference between a break statement and a continue Use Break Statement in jsp code after loop. break_statement_jsp.jsp <!DOCTYPE HTML PUBLIC... Use Break Statement in jsp code The break statement is used to terminate Java Break keyword . ?break? statement inside any loop bring the program control out of the loop. Example below demonstrates the working of break statement. In the program... Java Break keyword   Java Break command Java Break command Java Break command is commonly used in terminating looping statements. break command comes under the Java branching statements category. In programming Loop in java For Loop in Java Break statement in java Continue statement in Java... Loop is the control statement of any language in which whenever you want to perform the repetitious work then you use the Loop control statement adding loop a loop can you help me?The program below is compiling without the loop. >... the loop is: > Blockquote import java.util.*; class Output { public...(); if(code.equals("XX")){ break FOR Loop In JSP /DECREEMENT) { //loop statements; //'break;' statement can be used to exit...FOR Loop In JSP In this section we will read about how to use the for loop... are used e.g. while loop, do while loop, for loop. In programming these loops Java Break example . Java Break loop Java contains its own Branching... while etc. . Java break for loop Example..., a labeled break statement is used to terminate for loop.   loop problem - Java Magazine to main menu to select new item. //what loop to use? break...loop problem import java.util.Scanner; class mychoice{ public..., "A")); list.add(new Student(2, "B")); list.add(new Student(3, "C")); list.add how to intrupt for loopDarshan Manjunath July 8, 2013 at 12:40 AM hi, I want to intrupt a for loop which count till 30 or 60. I dont want to break the loop. If any intrupt(use of any key) occurs it should come out of the loop.. can anyone suggest me. Post your Comment
http://www.roseindia.net/discussion/23746-C-Break-for-loop.html
CC-MAIN-2015-06
refinedweb
1,115
76.32
import "text/tabwriter" Package tabwriter implements a write filter (tabwriter.Writer) that translates tabbed columns in input into properly aligned text. The package is using the Elastic Tabstops algorithm described at. The text/tabwriter package is frozen and is not accepting new features. //() Output: ....a|..b|c ...aa|.bb|cc ..aaa| .aaaa|.dddd|eeee //() Output: ------a|------b|---aligned| -----aa|-----bb|---aligned| ----aaa|----bbb|unaligned ---aaaa|---bbbb|---aligned|. A Writer is a filter that inserts padding around tab-delimited columns in its input to align them in the output. The Writer treats incoming bytes as UTF-8-encoded text consisting of cells terminated by horizontal ('\t') or vertical ('\v') tabs, and newline ('\n') or formfeed ('\f') characters; both newline and formfeed act as line breaks. Tab-terminated cells in contiguous lines constitute a column. The Writer inserts padding as needed to make all cells in a column have the same width, effectively aligning the columns. It assumes that all characters have the same width, except for tabs for which a tabwidth must be specified. Column cells must be tab-terminated, not tab-separated: non-tab terminated trailing text at the end of a line forms a cell but that cell is not part of an aligned column. For instance, in this example (where | stands for a horizontal tab): aaaa|bbb|d aa |b |dd a | aa |cccc|eee the b and c are in distinct columns (the b column is not contiguous all the way). The d and e are not in a column at all (there's no terminating tab, nor would the column be contiguous). The Writer assumes that all Unicode code points have the same width; this may not be true in some fonts or if the string contains combining characters. acts like a newline but it also terminates all columns in the current line (effectively calling Flush). Tab- terminated() Output: a b c d . 123 12345 1234567 123456789 . a b c d. 123 12345 1234567 123456789. Write writes buf to the writer b. The only errors returned are ones encountered while writing to the underlying output stream. Package tabwriter imports 2 packages (graph) and is imported by 9952 packages. Updated 2020-06-02. Refresh now. Tools for package owners.
https://godoc.org/text/tabwriter
CC-MAIN-2020-29
refinedweb
372
64.81
Drawing directly on a screen Every class that extends the Field class has a paint() method. The paint() method determines how a field draws its contents. For classes that are included in the BlackBerry Java. Context stack BlackBerry device applications that have a UI include one or more objects that extend the Screen class, and each of these objects represents a separate screen in the application. An application maintains one Graphics object per screen, and managers and other fields use this Graphics object to draw the contents of the fields. Each Graphics object contains an internal context stack that stores information about the area that a field should draw in. The context stack includes information such as clipping region (the boundaries of the area to draw in) and drawing offset (the position of the area to draw in). When a field is directed to draw itself and its contents, such as when the field is added to a screen or receives the focus, the field's manager pushes a transformation on to the context stack of the screen's Graphics object. The transformation provides information that determines how the field should be drawn on the screen. For example, the clipping region is defined as the dimensions of the field, so that the field can draw content only within the boundaries of the field. Similarly, the drawing offset is defined as the position of the field relative to the manager. The manager then passes the Graphics object, with its updated context stack, to the field that is being drawn. The field's paint() method uses the Graphics object to draw content in the proper position on the screen. A transformation is pushed on to the context stack using the coordinates of the parent object. For example, a manager pushes a transformation that contains coordinates relative to the manager, instead of relative to any of the fields that the manager contains. Thus, the paint() methods of each field can use local coordinates, with the upper-left corner of the field represented as (0, 0), when drawing. The coordinates are translated back up through the transformations in the context stack to appear correctly on the screen. You can push a new context on to the context stack of a Graphics object by invoking pushContext() of the Graphics class. Similarly, you can pop off the last pushed context stack by invoking popContext(). You can use these methods to create custom drawing areas on your screens. Drawing shapes You can use the methods of the Graphics class to draw primitive shapes on the screen of a BlackBerry device, including the following methods: - drawEllipse() - drawRect() - drawArc() These methods draw outlines of their respective shapes, without any fill. You can use corresponding fill methods (such as fillEllipse(), fillRect(), and fillArc()) to draw filled shapes. You can set properties for the shapes that you draw using the Graphics object, such as foreground color, background color, and alpha. Code sample: Drawing shapes on the screen import net.rim.device.api.ui.UiApplication; import net.rim.device.api.ui.container.MainScreen; import net.rim.device.api.ui.Graphics; public class DrawingShapesDemo extends UiApplication { public static void main(String[] args) { DrawingShapesDemo theApp = new DrawingShapesDemo(); theApp.enterEventDispatcher(); } public DrawingShapesDemo() { pushScreen(new DrawingShapesDemoScreen()); } } public class DrawingShapesDemoScreen extends MainScreen { public DrawingShapesDemoScreen() { setTitle("Drawing Shapes Demo"); } //Override paint() of the MainScreen class. protected void paint(Graphics g) { //To draw a colored rectangle: //Invoke setColor() to set the color of the rectangle. //Invoke fillRect() to draw a filled rectangle. //Invoke drawRect() to draw an empty rectangle. g.setColor(0x009900); g.fillRect(10, 10, 80, 80); g.drawRect(15, 15, 80, 80); //To draw a colored rectangle with rounded corners: invoke //setColor() to set the color of the rectangle. //Invoke fillRoundedRect() to draw a filled rectangle, and invoke //drawRoundedRect() to draw an empty rectangle. g.setColor(0x6699FF); g.fillRoundRect(120, 10, 80, 80, 25, 25); g.drawRoundRect(125, 15, 80, 80, 25, 25); //To draw a colored ellipse: invoke setColor() to set the //color of the ellipse. //Invoke fillEllipse() to draw a filled ellipse, and invoke //drawEllipse() to draw an empty ellipse. g.setColor(0xCCCCCC); g.fillEllipse(100, 200, 160, 240, 100, 270, 0, 360); g.drawEllipse(105, 205, 165, 245, 105, 275, 0, 360); } }
http://developer.blackberry.com/bbos/java/documentation/drawing_directly_on_a_screen_1970011_11.html
CC-MAIN-2014-15
refinedweb
708
54.12
core(4) core(4) core - format of core image file #include <core.out.h> The IRIX system writes out a core image of a terminated process when any of various errors occur. See signal(2) for the list of reasons; the most common are memory violations, illegal instructions, bus errors, and user-generated quit signals. The core image is called core and is written in the process's working directory (provided it can be; normal access controls apply). A process with an effective user ID different from the real user ID does not produce a core image. The format of the core image is defined by <core.out.h>. It consists of a header, maps, descriptors, and section data. The header data includes the process name (as in ps(1)), the signal that caused the core dump, the descriptor array, and the corefile location of the map array. Each descriptor defines the length of useful process data. One descriptor defines the general-purpose registers at the time of the core dump for example. The data is present in the core image at the file location given in the descriptor only if the IVALID flag is set in the descriptor. Each map defines the virtual address and length of a section of the process at the time of the core dump. The data is present in the core image at the file location given in the descriptor only if the VDUMPED flag is set in the map. The process's stack and data sections are normally written in the core image. The process's text is not normally written in the core image. Core image format designed by Silicon Graphics, Inc. dbx(1), ps(1), setuid(2), signal(2). PPPPaaaaggggeeee 1111
https://nixdoc.net/man-pages/IRIX/man4/core.4.html
CC-MAIN-2020-34
refinedweb
290
63.59
Pythonista for Secure Socket Connections? - Denrael245 I'm intrigued by Pythonista. I would love to use it for using my iPad to quickly do demos, but that would require that I be able to establish a secure socket connection. Does that capability exist as of yet? import ssl print(ssl.OPENSSL_VERSION) # 'OpenSSL 1.0.1b 26 Apr 2012' print(ssl.OPENSSL_VERSION_INFO) # (1, 0, 1, 2, 15) print(ssl.OPENSSL_VERSION_NUMBER) # 268439599L Also the HowTo at: Bummer! Our iOS devices are currently succeptable to the heartbleed bug in OpenSSL... For details, see: import ssl ; print(ssl.OPENSSL_VERSION) # 'OpenSSL 1.0.1b 26 Apr 2012' -- heartbleed is fixed in OpenSSL 1.0.1g and later I would expect Apple to release an iOS upgrade in the comming daze to raise the OpenSSL version to 1.0.1g or better but in the meantime, "secure" communications to/from your iOS device are not nearly as secure as you might have thought. If you used homebrew to install Python on your Mac, I would highly recommend that you again use homebrew to upgrade the OpenSSL that Python and Python3 use. This does not upgrade the system-wide OpenSSL (which is so old that it does not have heartbleed issues) but does upgrade the OpenSSL that homebrew-installed applications use.
https://forum.omz-software.com/topic/615/pythonista-for-secure-socket-connections
CC-MAIN-2017-47
refinedweb
214
76.82
How would I get all sub directories in the base master directory (C drive). I also need to use this for the registry (please I know what I am doing) How would I get all sub directories in the base master directory (C drive). I also need to use this for the registry (please I know what I am doing) well to bad the faq doesn't have any file recursion for C#. I'm sure the c++ code is convertable but I really don't know c++ as good as C#. p.s this is the C board not the C# board Last edited by prog-bman; 03-22-2005 at 07:16 PM. Woop? can someone move this please, I feel like an idoit. please move this to C# board. thank you for your patients with me. In other words, you can only get away with being a smart ass if you're not a dumb ass. Quzah. Hope is the first step on the road to disappointment. Moved to C# board as requested. However you may want to provide a little bit more detail. I want to collect the names of all of the directorys in the, lets say, C drive. From there I can get the names of the files, and any other info I need. That is as clear as I can get it. Simple recursion:Simple recursion: Originally Posted by Rune HunterOriginally Posted by Rune HunterCode:using System; using System.Collections; using System.IO; namespace GetDirs { class entry { static void Main(string[] args) { Test test = new Test(); test.GetDirs(@"C:\"); test.DumpDirs(); } } class Test { private ArrayList directories = new ArrayList(); public void GetDirs(string home) { string[] dirs1 = Directory.GetDirectories(home); foreach (string dir in dirs1) { directories.Add(dir); try { string[] dirs2 = Directory.GetDirectories(dir); if (dirs2.Length > 0) GetDirs(dir); } catch { } } } public void DumpDirs() { foreach (object obj in directories) Console.WriteLine((string)obj); } } } And here I modified the above example to do the same thing with registry keys:Code:using System; using System.Collections; using Microsoft.Win32; namespace GetReg { class entry { static void Main(string[] args) { Test test = new Test(); test.GetRegKeys(Registry.ClassesRoot); test.GetRegKeys(Registry.CurrentUser); test.GetRegKeys(Registry.LocalMachine); test.GetRegKeys(Registry.Users); test.GetRegKeys(Registry.CurrentConfig); test.DumpRegKeys(); } } class Test { private ArrayList regkeys = new ArrayList(); public void GetRegKeys(RegistryKey homekey) { string[] subkeys = homekey.GetSubKeyNames(); foreach (string subkey in subkeys) { try { regkeys.Add(homekey.Name + '\\' + subkey); RegistryKey subkey2 = homekey.OpenSubKey(subkey); if (subkey2.SubKeyCount > 0) GetRegKeys(subkey2); subkey2.Close(); } catch { } } } public void DumpRegKeys() { foreach (object obj in regkeys) Console.WriteLine((string)obj); } } } I don't see where test is coming from? It isn't in Visual C# 2005 at least. Edit: nvm I see now. omg thank you so much! I am going to study and learn how this works now! 1 problem, it says I don't have access to the registry but I am full adminster and have full permisions on this computer. Why can't I with .net? Edit: Hmm I guess my program I made before just was bad or somtin... it works fine now. Last edited by Rune Hunter; 03-24-2005 at 05:56 PM. ok so how would I get the value of the key it is in as well. I have tried many, many things but I got no results. + foreach + + foreach + Originally Posted by Rune HunterOriginally Posted by Rune Hunter Also, do a little search on recursion to understand why it works. alright I should do some research by just by looking at it I can grasp how it works. And thank you, I was alot confused on C# registry values. The registry can be confusing at times.
http://cboard.cprogramming.com/csharp-programming/63343-file-recursion-help.html
CC-MAIN-2016-36
refinedweb
615
68.36
Safari Books Online is a digital library providing on-demand subscription access to thousands of learning resources. In this chapter, you learned about streams and why they are used in the .NET Framework to access files and other serial devices. You looked at the basic classes in the System.IO namespace, including: You saw that the File class exposes many static methods for moving, copying, and deleting files, FileInfo represents a physical file on disk, and has methods to manipulate this file. A FileStream object represents a file that can be written to, or read from, or both. You also explored StreamReader and StreamWriter classes and saw how useful they were for writing to streams. You saw how to read and write to random files using the FileStream class. Building on this knowledge, you used classes in the System.IO.Compression namespace to compress streams as you write them to disk and also saw how to serialize objects to files. Finally, you built an entire application to monitor files and directories using the FileSystemWatcher class.
http://my.safaribooksonline.com/book/-/9780764578472/chapter-22-file-system-data/summary0038
CC-MAIN-2014-15
refinedweb
175
64.2
11 February 2010 10:54 [Source: ICIS news] SINGAPORE (ICIS news)--South Korea’s Yeochun Naphtha Cracking Center (YNCC) has reduced the operating rate of its 850,000 tonne/year No 1 naphtha cracker, located at Yeocheon, due to decoking activity, a company source said on Thursday. The company reduced the cracker’s run rate to 95% from 100%, the source said. It was not immediately clear when the decoking process would end, the source added. YNCC, the largest cracker operator in ?xml:namespace> On 20 January, a source from YNCC said the company would continue to operate its three crackers at maximum capacity in February amid strong petrochemical margins.
http://www.icis.com/Articles/2010/02/11/9333718/yncc-cuts-yeocheon-no-1-crackers-rate-to-95-due-to-decoking.html
CC-MAIN-2015-11
refinedweb
110
50.57
Problem: Who Fetched What, When? Running a couple of quick scripts over the logfile data reveals that there are 12,600,064 instances of an article fetch coming from 2,345,571 different hosts. Suppose we are interested in who was fetching what, and when? An auditor, a police officer, or a marketing professional might be interested. So, here’s the problem: given a hostname, report what articles were fetched from that host, and when. The result is a list; if the list is empty, no articles were fetched. We’ve already seen that a language’s built-in hash or equivalent data structure gives the programmer a quick and easy way to store and look up key/value pairs. So, you might ask, why not use it? That’s an excellent question, and we should give the idea a try. There are reasons to worry that it might not work very well, so in the back of our minds, we should be thinking of a Plan B. As you may recall if you’ve ever studied hash tables, in order to go fast, they need to have a small load factor; in other words, they need to be mostly empty. However, a hash table that holds 2.35 million entries and is still mostly empty is going to require the use of a whole lot of memory. To simplify things, I wrote a program that ran over all the logfiles and pulled out all the article fetches into a simple file; each line has the hostname, the time of the transaction, and the article name. Here are the first few lines: crawl-66-249-72-77.googlebot.com 1166406026 2003/04/08/Riff s egspd42470.ask.com 1166406027 2006/05/03/MARS-T-Shirt 84.7.249.205 1166406040 2003/03/27/Scanner (The second field, the 10-digit number, is the standard Unix/Linux representation of time as the number of seconds since the beginning of 1970.) Then I wrote a simple program to read this file and load a great big hash. Example 4-5 shows the program. EXAMPLE 4-5. Loading a big hash 1 class BigHash 2 3 def initialize(file) 4 @hash = {} 5 lines = 0 6 File.open(file).each_line do |line| 7 s = line.split 8 article = s[2].intern 9 if @hash[s[0]] 10 @hash[s[0]] << [ s[1], article ] 11 else 12 @hash[s[0]] = [ s[1], article ] 13 end 14 lines += 1 15 STDERR.puts "Line: #{lines}" if (lines % 100000) == 0 16 end 17 end 18 19 def find(key) 20 @hash[key] 21 end 22 23 end The program should be fairly self-explanatory, but line 15 is worth a note. When you’re running a big program that’s going to take a lot of time, it’s very disturbing when it works away silently, maybe for hours. What if something’s wrong? What if it’s going incredibly slow and will never finish? So, line 15 prints out a progress report after every 100,000 lines of input, which is reassuring. Running this program was interesting. It took about 55 minutes of CPU time to load up the hash, and the program grew to occupy 1.56 GB of memory. A little calculation sug gests that it costs around 680 bytes to store the information for each host, or slicing the data another way, about 126 bytes per fetch. This is a little scary, but probably reasonable for a hash table. Retrieval performance was excellent. I ran 2,000 queries, half of which were randomly selected hosts from the log and thus succeeded, while the other half were those same hostnames reversed, none of which succeeded. The 2,000 queries completed in an average of about .02 seconds, so Ruby’s hash implementation can look up records in a hash containing 12 million or so records thousands of times per second. Those 55 minutes to load up the data are troubling, but there are some tricks to address that. You could, for example, load it up once, then serialize the hash out and read it back in. And I didn’t try particularly hard to optimize the program. The program was easy and quick to write, and it runs fast once it’s initialized, so its performance is good both in terms of waiting-for-the-program time and waiting-for-the-programmer time. Still, I’m unsatisfied. I have the feeling that there ought to be a way to get this kind of performance while burning less memory, less startup time, or both. It involves writing our own search code, though. {mospagebreak title=Binary Search} Nobody gets a Computer Science degree without studying a wide variety of search algorithms: trees, heaps, hashes, lists, and more. My favorite among all these is binary search. Let’s try it on the who-fetched-what-when problem and then look at what makes it beautiful. My first attempt at putting binary search to use was quite disappointing; while the data took 10 minutes less to load, it required almost 100 MB more memory than with the hash. Clearly, there are some surprising things about the Ruby array implementation. The search also ran several times slower (but still in the range of thousands per second), but this is not surprising at all because the algorithm is running in Ruby code rather than with the underlying hardcoded hash implementation. The problem is that in Ruby everything is an object, and arrays are fairly abstracted things with lots of built-in magic. So, let’s reimplement the program in Java, in which integers are just integers, and arrays come with very few extras.* Nothing could be simpler, conceptually, than binary search. You divide your search space in two and see whether you should be looking in the top or bottom half; then you repeat the exercise until done. Instructively, there are a great many ways to code this algorithm incorrectly, and several widely published versions contain bugs. The implementation mentioned in “On the Goodness of Binary Search,” and shown in Java in Example 4-6, is based on one I learned from Gaston Gonnet, the lead developer of the Maple language for symbolic mathematics and currently Professor of Computer Science at ETH in Zürich. EXAMPLE 4-6. Binary search 1 package binary ; 2 3 public class Finder { 4 public static int find(String[] keys, String target) { 5 int high = keys.length; 6 int low = -1; 7 while (high – low > 1) { 8 int probe = (low + high) >>> 1; 9 if (keys[probe].compareTo(target) > 0) 10 high = probe; 11 else 12 low = probe; 13 } 14 if (low == -1 || keys[low].compareTo(target) != 0) 15 return -1; 16 else 17 return low; 18 } 19 } Key aspects of this program are as follows: - In lines 5–6, note that the high and low bounds are set one off the ends of the array, so neither are initially valid indices. This eliminates all sorts of corner cases. - The loop that starts in line 7 runs until the high and low bounds are adjacent; there is no testing to see whether the target has been found. Think for a minute whether you agree with this choice; we’ll return to the question later. The loop has two invariants. low is either –1 or points to something less than or equal to the target value. high is either one off the top of the array or points to something strictly greater than the target value. - Line 8 is particularly interesting. In an earlier version it read: probe = (high + low) / 2; but in June 2006, Java guru Josh Bloch showed how, in certain obscure circumstances, that code could lead to integer overflow (see extra-extra-read-all-about-it-nearly.html). It is sobering indeed that, many decades into the lifetime of computer science, we are still finding bugs in our core algorithms. (The issue is also discussed by Alberto Savoia in Chapter7.) At this point, Rubyists will point out that modern dynamic languages such as Ruby and Python take care of integer overflow for you, and thus don’t have this bug. - Because of the loop invariant, once I’m done with the loop, I just need to check low (lines 14–17). If it’s not –1, either it points to something that matches the target, or the target isn’t there. The Java version took only six and a half minutes to load, and it ran successfully, using less than 1 GB of heap. Also, while it’s harder to measure CPU time in Java than in Ruby, there was no perceptible delay in running the same 2,000 searches. {mospagebreak title=Binary Search Trade-offs} Binary search has some very large advantages. First of all, its performance is O(log2 N) . People often don’t really grasp how powerful this is. On a 32-bit computer, the biggest log2 you’ll ever encounter is 32 (similarly, 64 on a 64-bit computer), and any algorithm that competes in an upper bound of a few dozen steps will be “good enough” for many real-world scenarios. Second, the binary-search code is short and simple. Code that is short and simple is beau tiful, for a bunch of reasons. Maybe the most important is that it’s easier to understand, and understanding code is harder than writing it. There are fewer places for bugs to hide. Also, compact code plays better with instruction sets, I-caches, and JIT compilers, and thus tends to run faster. Third, once you’ve got that sorted array, you don’t need any more index structures; binary search is very space-efficient. The big downside to binary search is that the data has to be kept in order in memory. There are some data sets for which this is impossible, but fewer than you might think. If you think you have too much data to fit in memory, check the price of RAM these days and make sure. Any search strategy that requires going to disk is going to be immensely more complex, and in many scenarios slower. Suppose you need to update the data set; you might think that would rule out binary search because you have to update a huge, contiguous array in memory. But that turns out to be easier than you might think. In fact, your program’s memory is scattered randomly all over the computer’s physical RAM, with the operating system’s paging software making it look sequential; you can do the same kind of trick with your own data. Some might argue that since a hash table is O(1) , that has to be better than binary search’s O(log2 N ) . In practice, the difference may not be that significant; set up an experiment sometime and do some measurements. Also, consider that hash tables, with the necessary collision-resolution code, are considerably more complex to implement. I don’t want to be dogmatic, but in recent years, I’ve started to take the following approach to search problems: - Try to solve it using your language’s built-in hash tables. - Then try to solve it with binary search. - Only then should you reluctantly start to consider other more complex options. {mospagebreak title=Escaping the Loop} O(log2 N ) algorithm when the chances are it will save only a small number of steps at the end? The take-away lesson is that binary search, done properly, is a two-step process. First, write an efficient loop that positions your low and high bounds properly, then add a simple check to see whether you hit or missed. Search in the Large When.* Searching with Postings. Ranking Results. {mospagebreak title=Searching the Web} straightforward..
http://www.devshed.com/c/a/practices/more-techniques-for-finding-things/2/
CC-MAIN-2014-52
refinedweb
1,968
70.73
Basic features of FormFlow Dialogs are very powerful and flexible, but handling a guided conversation such as ordering a sandwich can require a lot of effort. At each point in the conversation, there are many possibilities of what will happen next. For example, you may need to clarify an ambiguity, provide help, go back, or show progress. By using FormFlow within the Bot Builder SDK for .NET, you can greatly simplify the process of managing a guided conversation like this. FormFlow automatically generates the dialogs that are necessary to manage a guided conversation, based upon guidelines that you specify. Although using FormFlow sacrifices some of the flexibility that you might otherwise get by creating and managing dialogs on your own, designing a guided conversation using FormFlow can significantly reduce the time it takes to develop your bot. Additionally, you may construct your bot using a combination of FormFlow-generated dialogs and other types of dialogs. For example, a FormFlow dialog may guide the user through the process of completing a form, while a LuisDialog may evaluate user input to determine intent. This article describes how to create a bot that uses the basic features of FormFlow to collect information from a user. Forms and fields To create a bot using FormFlow, you must specify the information that the bot needs to collect from the user. For example, if the bot's objective is to obtain a user's sandwich order, then you must define a form that contains fields for the data that the bot needs to fulfill the order. You can define the form by creating a C# class that contains one or more public properties to represent the data that the bot will collect from the user. Each property must be one of these data types: - Integral (sbyte, byte, short, ushort, int, uint, long, ulong) - Floating point (float, double) - String - DateTime - Enumeration - List of enumerations Any of the data types may be nullable, which you can use to model that the field does not have a value. If a form field is based on an enumeration property that is not nullable, the value 0 in the enumeration represents null (i.e., indicates that the field does not have a value), and you should start your enumeration values at 1. FormFlow ignores all other property types and methods. For complex objects, you must create a form for the top-level C# class and another form for the complex object. You can compose the forms together by using typical dialog semantics. It is also possible to define a form directly by implementing Advanced.IField or using Advanced.Field and populating the dictionaries within it. Note You can define a form by using either a C# class or JSON schema. This article describes how to define a form using a C# class. For more information about using JSON schema, see Define a form using JSON schema. Simple sandwich bot Consider this example of a simple sandwich bot that is designed to obtain a user's sandwich order. Create the form The SandwichOrder class defines the form and the enumerations define the options for building a sandwich. The class also includes the static BuildForm method that uses FormBuilder to create the form and define a simple welcome message. To use FormFlow, you must first import the Microsoft.Bot.Builder.FormFlow namespace. using Microsoft.Bot.Builder.FormFlow; using System; using System.Collections.Generic; // The SandwichOrder class represents the form that you want to complete // using information that is collected from the user. // It must be serializable so the bot can be stateless. // The order of fields defines the default sequence in which the user is asked questions. // The enumerations define the valid options for each field in SandwichOrder, and the order // of the values represents the sequence in which they are presented to the user in a conversation. namespace Microsoft.Bot.Sample.SimpleSandwichBot {(); } }; } Connect the form to the framework To connect the form to the framework, you must add it to the controller. In this example, the Conversation.SendAsync method calls the static MakeRootDialog method, which in turn, calls the FormDialog.FromForm method to create the SandwichOrder form. internal static IDialog<SandwichOrder> MakeRootDialog() { return Chain.From(() => FormDialog.FromForm(SandwichOrder.BuildForm)); } [ResponseType(typeof(void))] public virtual async Task<HttpResponseMessage> Post([FromBody] Activity activity) { if (activity != null) { switch (activity.GetActivityType()) { case ActivityTypes.Message: await Conversation.SendAsync(activity, MakeRootDialog); break; case ActivityTypes.ConversationUpdate: case ActivityTypes.ContactRelationUpdate: case ActivityTypes.Typing: case ActivityTypes.DeleteUserData: default: Trace.TraceError($"Unknown activity type ignored: {activity.GetActivityType()}"); break; } } ... } See it in action By simply defining the form with a C# class and connecting it to the framework, you have enabled FormFlow to automatically manage the conversation between bot and user. The example interactions shown below demonstrate the capabilities of a bot that is created by using the basic features of FormFlow. In each interaction, a > symbol indicates the point at which the user enters a response. Display the first prompt This form populates the SandwichOrder.Sandwich property. The form automatically generates the prompt, "Please select a sandwich", where the word "sandwich" in the prompt derives from the property name Sandwich. The SandwichOptions enumeration defines the choices that are presented to the user, with each enumeration value being automatically broken into words based upon changes in case and underscores. > Provide guidance The user can enter "help" at any point in the conversation to get guidance with filling out the form. For example, if the user enters "help" at the sandwich prompt, the bot will respond with this guidance. > help * You are filling in the sandwich field. Possible responses: * You can enter a number 1-15 or words from the descriptions. (BLT, Black Forest Ham, Buffalo Chicken, Chicken And Bacon Ranch Melt, Cold Cut Combo, Meatball Marinara, Oven Roasted Chicken, Roast Beef, Rotisserie Style Chicken, Spicy Italian, Steak And Cheese, Sweet Onion Teriyaki, Tuna, Turkey Breast, and Veggie) * Back: Go back to the previous question. * Help: Show the kinds of responses you can enter. * Quit: Quit the form without completing it. * Reset: Start over filling in the form. (With defaults from your previous entries.) * Status: Show your progress in filling in the form so far. * You can switch to another field by entering its name. (Sandwich, Length, Bread, Cheese, Toppings, and Sauce). Advance to the next prompt If the user enters "2" in response to the initial sandwich prompt, the bot then displays a prompt for the next property that is defined by the form: SandwichOrder.Length. > 2 Please select a length (1. Six Inch, 2. Foot Long) > Return to the previous prompt If the user enters "back" at this point in the conversation, the bot will return the previous prompt. The prompt shows the user's current choice ("Black Forest Ham"); the user may change that selection by entering a different number or confirm that selection by entering "c". > back Please select a sandwich(current choice: Black Forest Ham) > c Please select a length (1. Six Inch, 2. Foot Long) > Clarify user input If the user responds with text (instead of a number) to indicate a choice, the bot will automatically ask for clarification if user input matches more than one choice. Please select a bread 1. Nine Grain Wheat 2. Nine Grain Honey Oat 3. Italian 4. Italian Herbs And Cheese 5. Flatbread > nine grain By "nine grain" bread did you mean (1. Nine Grain Honey Oat, 2. Nine Grain Wheat) > 1 If user input does not directly match any of the valid choices, the bot will automatically prompt the user for clarification. Please select a cheese (1. American, 2. Monterey Cheddar, 3. Pepperjack) > amercan "amercan" is not a cheese option. > american smoked For cheese I understood American. "smoked" is not an option. If user input specifies multiple choices for a property and the bot does not understand any of the specified choices, it will automatically prompt the user for clarification. Please select one or more toppings 1. Banana Peppers 2. Cucumbers 3. Green Bell Peppers 4. Jalapenos 5. Lettuce 6. Olives 7. Pickles 8. Red Onion 9. Spinach 10. Tomatoes > peppers, lettuce and tomato By "peppers" toppings did you mean (1. Green Bell Peppers, 2. Banana Peppers) > 1 Show current status If the user enters "status" at any point in the order, the bot's response will indicate which values have already been specified and which values remain to be specified. Please select one or more sauce 1. Honey Mustard 2. Light Mayonnaise 3. Regular Mayonnaise 4. Mustard 5. Oil 6. Pepper 7. Ranch 8. Sweet Onion 9. Vinegar > status * Sandwich: Black Forest Ham * Length: Six Inch * Bread: Nine Grain Honey Oat * Cheese: American * Toppings: Lettuce, Tomatoes, and Green Bell Peppers * Sauce: Unspecified Confirm selections When the user completes the form, the bot will ask the user to confirm their selections. Please select one or more sauce 1. Honey Mustard 2. Light Mayonnaise 3. Regular Mayonnaise 4. Mustard 5. Oil 6. Pepper 7. Ranch 8. Sweet Onion 9. Vinegar > 1 Is this your selection? * Sandwich: Black Forest Ham * Length: Six Inch * Bread: Nine Grain Honey Oat * Cheese: American * Toppings: Lettuce, Tomatoes, and Green Bell Peppers * Sauce: Honey Mustard > If the user responds by entering "no", the bot allows the user to update any of the prior selections. If the user responds by entering "yes", the form has been completed and control is returned to the calling dialog. Is this your selection? * Sandwich: Black Forest Ham * Length: Six Inch * Bread: Nine Grain Honey Oat * Cheese: American * Toppings: Lettuce, Tomatoes, and Green Bell Peppers * Sauce: Honey Mustard > no What do you want to change? 1. Sandwich(Black Forest Ham) 2. Length(Six Inch) 3. Bread(Nine Grain Honey Oat) 4. Cheese(American) 5. Toppings(Lettuce, Tomatoes, and Green Bell Peppers) 6. Sauce(Honey Mustard) > 2 Please select a length (current choice: Six Inch) (1. Six Inch, 2. Foot Long) > 2 Is this your selection? * Sandwich: Black Forest Ham * Length: Foot Long * Bread: Nine Grain Honey Oat * Cheese: American * Toppings: Lettuce, Tomatoes, and Green Bell Peppers * Sauce: Honey Mustard > y Handling quit and exceptions If the user enters "quit" in the form or an exception occurs at some point in the conversation, your bot will need to know the step in which the event occurred, the state of the form when the event occurred, and which steps of the form were successfully completed prior to the event. The form returns this information via the FormCanceledException<T> class. This code example shows how to catch the exception and display a message according to the event that occurred. internal static IDialog<SandwichOrder> MakeRootDialog() { return Chain.From(() => FormDialog.FromForm(SandwichOrder.BuildLocalizedForm)) .Do(async (context, order) => { try { var completed = await order; // Actually process the sandwich order... await context.PostAsync("Processed your order!"); } catch (FormCanceledException<SandwichOrder> e) { string reply; if (e.InnerException == null) { reply = $"You quit on {e.Last} -- maybe you can finish next time!"; } else { reply = "Sorry, I've had a short circuit. Please try again."; } await context.PostAsync(reply); } }); } Summary This article has described how to use the basic features of FormFlow to create a bot that can: - Automatically generate and manage the conversation - Provide clear guidance and help - Understand both numbers and textual entries - Provide feedback to the user regarding what is understood and what is not - Ask clarifying questions when necessary - Allow the user to navigate between steps Although basic FormFlow functionality is sufficient in some cases, you should consider the potential benefits of incorporating some of the more advanced features of FormFlow into your bot. For more information, see Advanced features of FormFlow and Customize a form using FormBuilder. Sample code For complete samples that show how to implement FormFlow using the Bot Builder SDK for .NET, see the Multi-Dialog Bot sample and the Contoso Flowers Bot sample in GitHub. Next steps FormFlow simplifies dialog development. The advanced features of FormFlow let you customize how a FormFlow object behaves.
https://docs.microsoft.com/en-us/azure/bot-service/dotnet/bot-builder-dotnet-formflow?view=azure-bot-service-3.0
CC-MAIN-2018-30
refinedweb
1,986
57.06
/* * : tcl-p.c * purpose: check and return via exit code whether the tcl interface needs to be made * * $Header: /cvs/Darwin/Security/SecuritySNACCRuntime/tcl-p.c,v 1.1.1.1 2001/05/18 23:14:05 mb Exp $ * $Log: tcl-p.c,v $ * Revision 1.1.1.1 2001/05/18 23:14:05 mb * Move from private repository to open source repository * * Revision 1.2 2001/05/05 00:59:16 rmurphy * Adding darwin license headers * * Revision 1.1.1.1 1999/03/16 18:05:50 aram * Originals from SMIME Free Library. * * Revision 1.1 1995/07/25 22:24:48 rj * new file * */ #define COMPILER 1 #include "snacc.h" main() { #if TCL return 0; #else return 1; #endif }
http://opensource.apple.com//source/Security/Security-54/SecuritySNACCRuntime/tcl-p.c
CC-MAIN-2016-44
refinedweb
122
61.53
Arrrgh! This is hopeless!:mad: My dream is to be a game developer, and I have tried to make a simple 2D game. Every single time, I tell you, every single time, I get stuck with something. Before my imageIO hasn't been working, so I decided to make a game with only javas Graphics2D. Just when I got started I encounter a problem, my game loop doesn't work. I use a simple Thread game loop. I put print messages in my code to debug that way. The loop runs once and then stop! Why? Here is my code: Code : package shooter; import java.awt.Color; import java.awt.Graphics; import java.awt.Graphics2D; import java.awt.event.KeyAdapter; import java.awt.event.KeyEvent; import javax.swing.JPanel; public class GamePanel extends JPanel implements Runnable { private boolean paused = false; private boolean gameover = false; private int playerX, playerY, playerW, playerH, playerDx; private Thread thread; public GamePanel() { thread = new Thread(this); playerX = Game.WIDTH / 2 - 15; playerY = 360; playerW = 30; playerH = 20; addKeyListener(new KeyL()); thread.start(); } public void paint(Graphics gf) { Graphics2D g = (Graphics2D)gf; //paints the background. g.setColor(Color.BLACK); g.fillRect(0, 0, Game.WIDTH, Game.HEIGHT); //paints the player g.setColor(Color.WHITE); g.fillRect(playerX, playerY, playerW, playerH); System.out.println("painted!"); } //game loop @Override public void run() { move(); repaint(); System.out.println("looped!"); try { Thread.sleep(17); }catch(Exception ex){ System.out.println(ex); } } //checks all collisions in game public void checkCollisions() { //checks if the player its the walls. if (playerX + playerDx < 0) { playerX = 0; playerDx = 0; } else if (playerX + playerDx > Game.WIDTH) { playerX = Game.WIDTH - playerW; playerDx = 0; } } //moves all entities. public void move() { playerX += playerDx; System.out.println("Moved!"); } private class KeyL extends KeyAdapter { public void keyPressed(KeyEvent e) { int key = e.getKeyCode(); if (key == KeyEvent.VK_A) { playerDx = -2; System.out.println("moving left"); } if (key == KeyEvent.VK_D) { playerDx = 2; System.out.println("moving right"); } } } } Yes, I know there is no main class. That one is in the JFrame class but I know the error isn't there. This is very frustrating! Earlier I have just given up, but know I decided to get help instead. Please, please, PLEASE, help me! P.S. The controls doesn't work either.
http://www.javaprogrammingforums.com/%20whats-wrong-my-code/16348-help-me-printingthethread.html
CC-MAIN-2014-35
refinedweb
377
63.15
Re: type conversion error - From: Paolo <Paolo@xxxxxxxxxxxxxxxxxxxxxxxxx> - Date: Fri, 29 Feb 2008 05:24:01 -0800 Hi neerak, Take a look to the missing data in your csv file so you can understand why they are discarded. Perhaps they are more than 255 char so they can't be contained in a text field so you must define it as memo. HTH Paolo "neerak via AccessMonster.com" wrote: Thanks for Allen for your suggestion, but i've almost tried every single. method mentioned in this forum and it still doesn't work. I've tried to convert the columns in the csv file to text before the transfertext method as follows xlwkbk.worksheets(1).columns("A").numberformat = "@" DoCmd.TransferText acImportDelim, , "tmpTable", CurrentProject.Path & "\test. csv", no and it doesn't work. then i tried to create tmpTable with all text fields and insert "ABC" in the first line of every column in the csv file then execute transfertext... but still it doesn't work. These two methods will generate the type conversion error table after execution. Then i moved on to use the following function BuildJetTextSource written by John Nurick with the following statements and this time I didn't get the type conversion error table but some of the values in the columns (the ones that generate the conversion error previously) were missing in the target table. strSQL = "insert into targetTable (ACCOUNT_ID,DEPT_NAME) SELECT f1,f2 from " & _ 'BuildJetTextSource("C:\test.csv", False) _ '& ";" Function BuildJetTextSource(ByVal FileSpec As String, _ ByVal HDR As Boolean) As String 'Takes a filespec and returns a string that can be used 'in the FROM clause of a Jet SQL query. ' E.g. ' C:\My Folder\MyFile.txt ' returns ' [Text;HDR=No;Database=C:\My Folder\;].[MyFile#txt] ' The HDR argument controls the HDR parameter in string returned. ' ' By John Nurick 2005 ' Revised 2007 to remove call to Dir() Dim fso As Object 'FileSystemObject Dim strFolder As String Dim strFileName As String Dim strFileExt As String Dim strTemp As String 'Parse FileSpec Set fso = CreateObject("Scripting.FileSystemObject") With fso FileSpec = .GetAbsolutePathName(FileSpec) strFolder = .GetParentFolderName(FileSpec) strFileName = .GetBaseName(FileSpec) strFileExt = .GetExtensionName(FileSpec) End With Set fso = Nothing 'Build string strTemp = "[Text;HDR=" _ & IIf(HDR, "Yes", "No") _ & ";Database=" _ & strFolder & "\;].[" _ & strFileName & "#" _ & strFileExt & "]" BuildJetTextSource = strTemp End Function Can anyone please help me with this problem? Or is there a way such that i can programatically execute the import method in the menu bar. Because manually importing the same csv file does not encounter any errors. Thank you.... Allen Browne wrote: The safest way to do this is to create another table with Text fields (not numbers or dates or curreny), and use an Append query to populate this table from the text file. Since they are Text fields, Access can stick the data in. You can then massage the data to populate your real table with the correct data types. I kept on receiving type conversion error while trying to programatically[quoted text clipped - 20 lines] import a csv file into an access table. I've read a lot of the posts such that the error only occurs when i tried to import the data programatically? Any help will be much appreciated.... -- Message posted via AccessMonster.com - References: - type conversion error - From: neerak via AccessMonster.com - Re: type conversion error - From: Allen Browne - Re: type conversion error - From: neerak via AccessMonster.com - Prev by Date: Re: Help needed on some code. - Next by Date: RE: Printing Sub Reports in Access - Previous by thread: Re: type conversion error - Next by thread: Re: type conversion error - Index(es):
http://www.tech-archive.net/Archive/Access/microsoft.public.access.modulesdaovba/2008-02/msg00867.html
crawl-002
refinedweb
605
54.42
First Steps¶ Warning The current page still doesn't have a translation for this language. But you can help translating it: Contributing. The simplest FastAPI file could look like this: from fastapi import FastAPI app = FastAPI() @app.get("/") async def root(): return {"message": "Hello World"} Copy that to a file main.py. Run the live server: $. Note The command uvicorn main:app refers to: main: the file main.py(the Python "module"). app: the object created inside of main.pywith the line app = FastAPI(). --reload: make the server restart after code changes. Only use for development. In the output, there's a line with something like: INFO: Uvicorn running on (Press CTRL+C to quit) That line shows the URL where your app is being served, in your local machine. Check it¶ Open your browser at. You will see the JSON response as: {"message": "Hello World"} Interactive API docs¶ Now go to. You will see the automatic interactive API documentation (provided by Swagger UI): Alternative API docs¶ And now, go to. You will see the alternative automatic documentation (provided by ReDoc): OpenAPI¶ FastAPI generates a "schema" with all your API using the OpenAPI standard for defining APIs. "Schema"¶ A "schema" is a definition or description of something. Not the code that implements it, but just an abstract description. API "schema"¶ In this case, OpenAPI is a specification that dictates how to define a schema of your API. This schema definition includes your API paths, the possible parameters they take, etc. Data "schema"¶ The term "schema" might also refer to the shape of some data, like a JSON content. In that case, it would mean the JSON attributes, and data types they have, etc. OpenAPI and JSON Schema¶ OpenAPI defines an API schema for your API. And that schema includes definitions (or "schemas") of the data sent and received by your API using JSON Schema, the standard for JSON data schemas. Check the openapi.json¶ If you are curious about how the raw OpenAPI schema looks like, FastAPI automatically generates a JSON (schema) with the descriptions of all your API. You can see it directly at:. It will show a JSON starting with something like: { "openapi": "3.0.2", "info": { "title": "FastAPI", "version": "0.1.0" }, "paths": { "/items/": { "get": { "responses": { "200": { "description": "Successful Response", "content": { "application/json": { ... What is OpenAPI for¶ The OpenAPI schema is what powers the two interactive documentation systems included. And there are dozens of alternatives, all based on OpenAPI. You could easily add any of those alternatives to your application built with FastAPI. You could also use it to generate code automatically, for clients that communicate with your API. For example, frontend, mobile or IoT applications. Recap, step by step¶ Step 1: import FastAPI¶ from fastapi import FastAPI app = FastAPI() @app.get("/") async def root(): return {"message": "Hello World"} FastAPI is a Python class that provides all the functionality for your API. Technical Details FastAPI is a class that inherits directly from Starlette. You can use all the Starlette functionality with FastAPI too. Step 2: create a FastAPI "instance"¶ from fastapi import FastAPI app = FastAPI() @app.get("/") async def root(): return {"message": "Hello World"} Here the app variable will be an "instance" of the class FastAPI. This will be the main point of interaction to create all your API. This app is the same one referred by uvicorn in the command: $ uvicorn main:app --reload <span style="color: green;">INFO</span>: Uvicorn running on (Press CTRL+C to quit) If you create your app like: from fastapi import FastAPI my_awesome_api = FastAPI() @my_awesome_api.get("/") async def root(): return {"message": "Hello World"} And put it in a file main.py, then you would call uvicorn like: $ uvicorn main:my_awesome_api --reload <span style="color: green;">INFO</span>: Uvicorn running on (Press CTRL+C to quit) Step 3: create a path operation¶ Path¶ "Path" here refers to the last part of the URL starting from the first /. So, in a URL like: ...the path would be: /items/foo Info A "path" is also commonly called an "endpoint" or a "route". While building an API, the "path" is the main way to separate "concerns" and "resources". Operation¶ "Operation" here refers to one of the HTTP "methods". One of: GET PUT DELETE ...and the more exotic ones: OPTIONS HEAD PATCH TRACE In the HTTP protocol, you can communicate to each path using one (or more) of these "methods". When building APIs, you normally use these specific HTTP methods to perform a specific action. Normally you use: POST: to create data. GET: to read data. PUT: to update data. DELETE: to delete data. So, in OpenAPI, each of the HTTP methods is called an "operation". We are going to call them "operations" too. Define a path operation decorator¶ from fastapi import FastAPI app = FastAPI() @app.get("/") async def root(): return {"message": "Hello World"} The @app.get("/") tells FastAPI that the function right below is in charge of handling requests that go to: - the path / - using a getoperation FastAPI that the function below corresponds to the path / with an operation get. It is the "path operation decorator". You can also use the other operations: @app.post() @app.put() @app.delete() And the more exotic ones: @app.options() @app.head() @app.patch() @app.trace() Tip You are free to use each operation (HTTP method) as you wish. FastAPI doesn't enforce any specific meaning. The information here is presented as a guideline, not a requirement. For example, when using GraphQL you normally perform all the actions using only POST operations. Step 4: define the path operation function¶ This is our "path operation function": - path: is /. - operation: is get. - function: is the function below the "decorator" (below @app.get("/")). from fastapi import FastAPI app = FastAPI() @app.get("/") async def root(): return {"message": "Hello World"} This is a Python function. It will be called by FastAPI whenever it receives a request to the URL " /" using a GET operation. In this case, it is an async function. You could also define it as a normal function instead of async def: from fastapi import FastAPI app = FastAPI() @app.get("/") def root(): return {"message": "Hello World"} Note If you don't know the difference, check the Async: "In a hurry?". Step 5: return the content¶ from fastapi import FastAPI app = FastAPI() @app.get("/") async def root(): return {"message": "Hello World"} You can return a dict, list, singular values as str, int, etc. You can also return Pydantic models (you'll see more about that later). There are many other objects and models that will be automatically converted to JSON (including ORMs, etc). Try using your favorite ones, it's highly probable that they are already supported. Recap¶ - Import FastAPI. - Create an appinstance. - Write a path operation decorator (like @app.get("/")). - Write a path operation function (like def root(): ...above). - Run the development server (like uvicorn main:app --reload).
https://fastapi.tiangolo.com/tr/tutorial/first-steps/
CC-MAIN-2021-17
refinedweb
1,147
67.15
Before we get into this tutorial we need to define what a closure is. The Camel (3rd edition) states that a closure is when you define an anonymous function in a particular lexical scope at any particular moment Now with that (simple?) definition out of the way, we can get on with the show! [download]: [] [download]] [download]! [download]! [download] So if we further apply the concepts of closures we can write ourselves a very basic directory iterator [download] Wrapping it up This method of creating closures using anonymous subroutines can be very powerful[1]. With the help of Richard Clamp's marvellous File::Find::Rule we can build ourselves a handy little grep like tool for XML files exi +t: } [download] _________broquaint Disclaimer: Not everyone will agree with the terminology (I imagine) so as long as you don't find the descriptions wildly off the mark or generally misleading then they're likely to stay as they are. Update - revised the second and third sections by dropping any references to 'reference counting' ++ prefix Object is data wrapped in methods. Closure is subroutine wrapped in data. use strict; use warnings; use Benchmark qw(cmpthese); # - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - sub closure_counter { my $count = shift; return sub { $count++ }; } # - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - {package obj_counter; sub new {bless {count=>$_[1]};} sub count {return ($_[0]->{count})++} } # - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - my $obj_counter = obj_counter->new(3); my $closure_counter = closure_counter(3); cmpthese(1000, { obj_counter=>sub{$obj_counter->count()for 1..1000;}, closure_counter=>sub{$closure_counter->()for 1..1000;} } ); __END__ Benchmark: timing 1000 iterations of closure_counter, obj_counter... closure_counter: 2 wallclock secs ( 2.58 usr + 0.00 sys = 2.58 CPU) + @ 387.60/s (n=1000) obj_counter: 5 wallclock secs ( 5.22 usr + 0.00 sys = 5.22 CPU) @ 1 +91.57/s (n=1000) Rate obj_counter closure_counter obj_counter 192/s -- -51% closure_counter 388/s 102% -- [download] use strict; use warnings; use Benchmark qw(cmpthese); # - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - use IO::Dir; sub closure_dir_iter { my $dir = IO::Dir->new(shift); return sub { my $fl = $dir->read(); $dir->rewind() unless defined $fl; return $fl; }; } # - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - {package obj_dir_iter; use IO::Dir; sub new {bless {dir=>IO::Dir->new($_[1])};} sub iter { my $fl = $_[0]->{dir}->read(); $_[0]->{dir}->rewind() unless defined $fl; return $fl; } } # - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - my $obj_dir_iter = obj_dir_iter->new( "." ); my $closure_dir_iter = closure_dir_iter( "." ); cmpthese(500, { obj_dir_iter=>sub{while(defined(my $f = $obj_dir_iter->iter()) +){print "$f\n";}}, closure_dir_iter=>sub{while(defined(my $f = $closure_dir_iter- +>())){print "$f\n";}} } ); __END__ Benchmark: timing 2000 iterations of closure_dir_iter, obj_dir_iter... obj_dir_iter: 1 wallclock secs ( 1.20 usr + 0.00 sys = 1.20 CPU) @ +1666.67/s (n=2000) closure_dir_iter: 1 wallclock secs ( 1.10 usr + 0.00 sys = 1.10 CPU +) @ 1818.18/s (n=2000) Rate obj_dir_iter closure_dir_iter obj_dir_iter 1667/s -- -8% closure_dir_iter 1818/s 9% -- [download] use strict; use warnings; use Benchmark qw(cmpthese); # - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - use constant PI => 3.14159265359; sub closure_turtle { my ($h, $xy) = (0, [[0],[0]]); # h = heading (0 - north, 90 - +east, etc) return sub { $h = $h + (shift || 0); # accumulative turns in degree my $d = shift || 0; # distance $xy->[0][scalar(@{$xy->[0]})] = $d*sin(PI*$h/180) + $xy->[0][$ +#{@{$xy->[0]}}]; $xy->[1][scalar(@{$xy->[1]})] = $d*cos(PI*$h/180) + $xy->[1][$ +#{@{$xy->[1]}}]; return $xy; }; } sub closure_koch { my ($turtle, $d, $level) = @_ ; if ($level==0) {$turtle->(0,$d); return 1;} $turtle->( 0,0); closure_koch($turtle,$d/3,$level-1); $turtle->(-60,0); closure_koch($turtle,$d/3,$level-1); $turtle->(120,0); closure_koch($turtle,$d/3,$level-1); $turtle->(-60,0); closure_koch($turtle,$d/3,$level-1); } # - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - {package obj_turtle; use constant PI => 3.14159265359; sub new {return bless({h=>0,xy=>[[0],[0]]});} sub rt {$_[0]->{h}=$_[0]->{h}+$_[1];} # right turn by x + degrees sub fd { # forward by x poi +nts my ($h, $xy, $d) = ($_[0]->{h}, $_[0]->{xy}, $_[1]); $xy->[0][scalar(@{$xy->[0]})] = $d*sin(PI*$h/180) + $xy->[0][$ +#{@{$xy->[0]}}]; $xy->[1][scalar(@{$xy->[1]})] = $d*cos(PI*$h/180) + $xy->[1][$ +#{@{$xy->[1]}}]; $_[0]->{xy} = $xy; return $xy; } } sub obj_koch { my ($turtle, $d, $level) = @_ ; if ($level==0) {$turtle->fd($d); return 1;} $turtle->rt( 0); obj_koch($turtle, $d/3,$level-1); $turtle->rt(-60); obj_koch($turtle, $d/3,$level-1); $turtle->rt(120); obj_koch($turtle, $d/3,$level-1); $turtle->rt(-60); obj_koch($turtle, $d/3,$level-1); } # - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - my $obj_turtle = obj_turtle->new(); my $closure_turtle = closure_turtle(); cmpthese(100, { obj_turtle=>sub{for(0..2){$obj_turtle->rt(120); obj_koch($obj_ +turtle,170,4);}}, closure_turtle=>sub{for(0..2){$closure_turtle->(120, 0); closu +re_koch($closure_turtle,170,4);}} } ); __END__ Benchmark: timing 100 iterations of closure_turtle, obj_turtle... closure_turtle: 6 wallclock secs ( 6.49 usr + 0.00 sys = 6.49 CPU) +@ 15.41/s (n=100) obj_turtle: 5 wallclock secs ( 4.99 usr + 0.00 sys = 4.99 CPU) @ 20 +.04/s (n=100) Rate closure_turtle obj_turtle closure_turtle 15.4/s -- -23% obj_turtle 20.0/s 30% -- [download] # IF = %b.(%x.(%y.(b x) y)) $IF = sub { my $b = shift; sub { my $x = shift; sub { my $y = shift; $b->($x)->($y); } } } # TRUE = %x.(%y.x) $TRUE = sub { my $x = shift; sub { my $y = shift; $x; } } # FALSE = %x.(%y.y) $FALSE = sub { my $x = shift; sub { my $y = shift; $y; } } print $IF->($TRUE)->("then")->("else"); # prints "then" print $IF->($FALSE)->("then")->("else"); # prints "else" [download] (IF TRUE A B) (%b.%x.%y.(b x y) %a.%b.a A B) (%x.%y.(%a.%b.a x y) A B) (%y.(%a.%b.a A y) B) (%a.%b.a A B) (%b.A B) A. However,."; I. I
http://www.perlmonks.org/?node_id=268891
CC-MAIN-2015-48
refinedweb
906
68.16
Hello all! It all started the night before Thanksgiving. It was 11:58 PM, and I was up late doing pretty much nothing when I had an idea… “What if I build a robot to pass food around the table?” So I stayed up all night long to build it, which in reality took me around 2 total hours, and the other 8 were spent dozing in and out because staying up was a terrible idea like this entire thing was. The plan was simple: use 2 motors with 3D printed brackets, a paper plate, an Arduino Nano, and an ultrasonic sensor to make a robot that moves around the table. When the robots distance to an object in front of it is greater than 5 centimeters, but less than 150 centimeters, it will move forwards until it is less than 5, at which point it will stop moving and wait 30 seconds before turning. If the distance is greater than 150 cm it will also turn, as this generally assumes that it has reached the edge of the table and it may fall, so it will turn away from it. Arduino is written in C++, so there are 2 main functions in the code, as well as the ability to declare global variables outside of these functions. We will begin by declaring variables that hold the values for which pins we want to use as our motor outputs, our ultrasonic sensor input and output, include the ultrasonic sensor library, and instantiate the ultrasonic sensor. #include <HCSR04.h> int triggerPin = 6; int echoPin = 7; #define m1low 2 #define m1high 3 #define m2low 4 #define m2high 5 UltraSonicDistanceSensor distanceSensor(triggerPin, echoPin); // M1 is motor 1, which is on the left side // M2 is motor 2, which is on the right side The next function is called the setup function, which is generally used to instantiate classes and setup things like serial monitors. In our case, we are using it to assign the pinModes for our different input and output pins, as well as set up a serial connection for the ultrasonic sensor. void setup() { Serial.begin(9600); pinMode(m1low, OUTPUT); pinMode(m1high, OUTPUT); pinMode(m2low, OUTPUT); pinMode(m2high, OUTPUT); } After we’ve setup all of our pins to be used, we setup a few different loops to determine the distance between the robot and the nearest object. As described above, if this distance is greater than 150cm then the robot will turn, if it is between 150cm and 5cm it will go straight forwards, and if it is less than 5 it will turn. We can grab the distance using the .measureDistanceCm(); method from ultrasonic sensor library. void loop() { double distance = distanceSensor.measureDistanceCm(); Serial.println(distance); while(distance >= 150){ digitalWrite(m1high, HIGH); digitalWrite(m2low, HIGH); digitalWrite(m1low, LOW); digitalWrite(m2high, LOW); distance = distanceSensor.measureDistanceCm(); } if(distance >= 5 && distance < 150){ digitalWrite(m1low, LOW); digitalWrite(m2low, LOW); digitalWrite(m1high, HIGH); digitalWrite(m2high, HIGH); delay(500); } if(distance <= 5){ digitalWrite(m1high, HIGH); digitalWrite(m2low, HIGH); digitalWrite(m1low, LOW); digitalWrite(m2high, LOW); delay(30000); digitalWrite(m1low, LOW); digitalWrite(m2low, LOW); digitalWrite(m1high, LOW); digitalWrite(m2high, LOW); } } And like that we have finished the code! Now onto the physical build. I started by designing a bracket for the motors I had, this bracket is up on my Thingiverse and in the GitHub repo for this project. I originally planned to screw these into a circle of wood (hence the holes for screws) but never ended up doing this and instead hot glued it to a paper plate. Next up I pulled out a breadboard and an Arduino nano. I wired pins 2 and 3 to the first motor, with a 1K ohm resistor on pin 2 so I could have a fast and high speed. I did this the same for pins 4 and 5 for the second motor. Then I placed pins 6 and 7 as the trigger and echo for the ultrasonic sensor. I finished up the electronics by using a spare 9 volt to 5 volt board I had to power the Arduino, and duck taping it all very nicely to the bottom of the plate. Here are some photos from the process: Overall the robot was a ‘success’ in the sense that it was just as terrible as I expected it to be. The reaction from my family members was hilarious as I pulled the robot out and placed it on the table, seeing the confusion and impending doom as I turned it on and it inched forwards before dropping a piece of pie onto the table. It was also amazingly hilarious that the robot worked better upside down than the way it was originally designed to work in. Thanks for reading and have a wonderful day! ~ Corbin
https://maker.godshell.com/archives/277
CC-MAIN-2020-45
refinedweb
796
54.15
How to Build a Soundboard with JavaScript August 27th, 2021 What You Will Learn in This Tutorial How to build a soundboard in JavaScript by creating a SoundPlayer class that dynamically injects players and makes it easy to map their playback to a DOM event. Table of Contents Master Websockets — Learn how to build a scalable websockets implementation and interactive UI. Getting Started For this tutorial, we're going to use the CheatCode Next.js Boilerplate as a starting point for our work. To start, let's clone a copy: Terminal git clone cd into the project and install its dependencies: Terminal cd nextjs-boilerplate && npm install Finally, start up the development server: Terminal npm run dev With all of that, we're ready to get started. Building a sound player In order to actually play the sounds in our soundboard, we'll want an easy way to create audio players on-the-fly. To do that, we're going to begin by wiring up a JavaScript class that will handle the creation of the <audio></audio> elements that will play our sounds and automate the injection of those elements into the DOM. /lib/soundPlayer.js class SoundPlayer { constructor() { this.sounds = []; } // We'll implement the API for our class here... } export default SoundPlayer; To start, here, we're creating a skeleton for our SoundPlayer class which will help us to load sounds into the DOM as well as play those sounds. Here, we set up a basic JavaScript class and export it as the default from /lib/soundPlayer.js. Inside the class, we add the constructor function (this is what gets called right as our class is loaded into memory by JavaScript) and initialize the sounds property on the class, setting it to an empty this is referring to the current class instance of SoundPlayer. We're creating an array here as we'll want a way to keep track of all the sounds we've loaded into the DOM. []array. Here, Terminal); } } export default SoundPlayer; Next, we need a simple API (application programming interface, here used colloquially to mean "the implementation of the player") for loading sounds into the DOM. To do it, above, we're adding two methods to our class: load() and injectPlayerIntoPage(). The first will be a publically-exposed function that we'll call from our UI to say "load this sound into the DOM." Inside of that function, we can see two things happening. First, like we hinted at above, we want to keep track of the sounds we're loading in. Taking in a name argument (an easy-to-remember name to "label" our sound by) and a path (the literal path to the sound file in our app), we overwrite the this.sounds property on our class to be equal to the current value of this.sounds, concatenated with a new object containing the name and path passed into load(). Here, ...this.sounds is "unpacking" the entirety of the existing this.sounds array (whether or not it contains anything). The ...part is known as the spread operator in JavaScript (it "spreads out" the contents of the value immediatelly following the ...). Next, with our this.sounds array updated, we need to dynamically create the <audio></audio> element we talked about above. To do it, we're adding a separate method injectPlayerIntoPage() which takes in the same two arguments from load(), name and path. Inside of that function, the first thing we need to do is create the <audio></audio> element in memory. To do it, we run document.createElement('audio') to instruct JavaScript to create an in-memory (meaning not added to the screen/DOM yet) copy of our <audio></audio> element. We store the result of that (the in-memory DOM node for our <audio></audio> element) in the variable const player. We do this to more easily modify the attributes of the player and then append it to the DOM. Specifically, we set four properties to our player before we append it to the DOM: idwhich is set to the namewe passed in for our sound. srcwhich is set to the pathto the file on the computer for the sound. volumewhich is set to 0.5or 50% to ensure we don't shatter our user's ear drums. typewhich is set to the file type we expect for our files (for our example, we're using .mp3files so we used the audio/mpegMIME-type-find others here). Once we've set all of these properties, finally, we use appendChild on document.body to append our audio player to the DOM (the physical location of this in the DOM is irrelevant as we'll learn next). /lib/soundPlayer.js); } play(name) { const player = document.getElementById(name); if (player) { player.pause(); player.currentTime = 0; player.play(); } } } export default SoundPlayer; To wrap up our SoundPlayer class, we need to add one more method: play(). Like the name suggests, this will play a sound for us. To do it, first, we take in a name argument (one that we would have passed into load() earlier) and try to find an element on the page with an id attribute matching that name. Recall that above we set the .id on our <audio></audio> tag to the name we passed in. This should find a match in the DOM. If it does, we first .pause() the player (in case we're mid-playback already), force the .currentTime attribute on the player to 0 (i.e., the start of our sound), and then .play() it. That does it for our SoundPlayer class. Next, let's wire it up and start to play some sounds! Adding a React page component to test our player Because our boilerplate is based on Next.js, now, we're going to create a new page in our app using a React.js component where we can test out our SoundPlayer. /pages/soundboard/index.js import React from "react"; import SoundPlayer from "../../lib/soundPlayer"; class Soundboard extends React.Component { state = { sounds: [ { name: "Kick", file: "/sounds/kick.mp3" }, { name: "Snare", file: "/sounds/snare.mp3" }, { name: "Hi-Hat", file: "/sounds/hihat.mp3" }, { name: "Tom", file: "/sounds/tom.mp3" }, { name: "Crash", file: "/sounds/crash.mp3" }, ], }; componentDidMount() { const { sounds } = this.state; this.player = new SoundPlayer(); sounds.forEach(({ name, file }) => { this.player.load(name, file); }); } render() { const { sounds } = this.state; return ( <div> {sounds.map(({ name, file }) => { return ( <button className="btn btn-primary btn-lg" style={{ marginRight: "15px" }} onClick={() => this.player.play(name)} > {name} </button> ); })} </div> ); } } Soundboard.propTypes = {}; export default Soundboard; In Next.js, routes or URLs in our app are automatically created by the the framework based on the contents of the /pages folder at the root of our app. Here, to create the route /soundboard (this will ultimately be accessible via in the browser), we create the folder /pages/soundboard and put an index.js file in that folder where the React component representing our page will live. Because our test component is so simple, above, we've output the entire contents. Let's step through it to understand how all of this is fitting together. First, up top we import our SoundPlayer class from our /lib/soundPlayer.js file. Next, we define a React component using the class-based method (this makes it easier to work with our player and avoid performance issues). The first part we want to call attention to is the state property we're adding to the class and the sounds property we've set to an array of objects there. This should be starting to make some sense. Here, we're creating all of the sounds that we want to load into the DOM using the load() method we wrote earlier on our SoundPlayer class. Remember, that function takes a name and a file argument which we're defining here. We do this as an array of objects to make it easier to loop over and load all of our sounds at once, which we do in the componentDidMount() function on our React component. In there, we use JavaScript object destructuring to "pluck off" the sounds property we just defined on state (accessible in our component's methods as this.state) and then create an instance of our SoundPlayer class with new SoundPlayer() and then assign that instance back to this.player on our Soundboard component class (this will come in handy soon). Next, using that sounds array we defined on state, we loop over it with a .forEach(), again using JavaScript destructuring to "pluck off" the name and file properties of each object in the array as we loop over them. With these values, we call to this.player.load(), passing them into the function. Like we learned earlier, we expect this to add each of the sounds in our array to the this.sounds array on our SoundPlayer class' instance and then append a DOM element for that sound's <audio></audio> player. Where this all comes together is down in the render() method on our component class. Here, we again "pluck off" the sounds array from this.state, this time using a JavaScript .map() to loop over the array, allowing us to return some markup that we want React to render for each iteration (each sound) of our array. Because we're building a soundboard, we add a <button></button> for each sound with an onClick attribute set to a function which calls this.player.play() passing in the name attribute from the sound's object in the this.state.sounds array. With this, we have a soundboard! Now when we click on a button, we should hear the associated sound in the file play back. Download a .zip containing all of the sounds referenced in the code above and unzip its contents into your app's /public/soundsfolder from CheatCode's Amazon S3 Bucket. That's it! If you'd like to add your own custom sounds, just make sure to add them to the /public/sounds folder in your app and then update the sounds array on state. Wrapping up In this tutorial, we learned how to create a soundboard using JavaScript. To do it, we began by creating a JavaScript class that helped us to dynamically create audio players that we could reference by a unique name. On that class, we also added a .play() method to streamline the playback of our sounds. To build the UI for our soundboard, we defined a React component that created an instance of our soundboard class, loaded in our preferred list of sounds, and then rendered a list of buttons, each with a call to the .play() method for the sound represented by that button. Get the latest free JavaScript and Node.js tutorials, course announcements, and updates from CheatCode in your inbox. No spam. Just new tutorials, course announcements, and updates from CheatCode.
https://cheatcode.co/tutorials/how-to-build-a-soundboard-with-javascript
CC-MAIN-2022-21
refinedweb
1,814
73.27
? Delegates are at the core of a number of different .NET facilities, events in particular. It’s long been a truism that the way to get work done is to delegate, but what are C# delegates all about? First we have to understand what the problem is that delegates are designed to solve. In many languages functions are treated in the same way as objects. In JavaScript for example function are objects - this is often expressed by saying the functions are first class objects or they are first class functions. Why should this matter? The simple answer is that sometimes you want to pass a function as an argument to a method call. There are other things you might want to do with a first class function including having a reference to it or having an array of such functions but if we concentrate on the most common usage - passing functions to methods - then we have the key idea. For example suppose you have a sorting routine you might want to pass in a function that determines what A>B means, i.e. return true if A is greater than B and false otherwise, where A and B are objects of some type. This seems like a reasonable thing to do but notice that in C# and in other object oriented language there are no "disembodied" functions. All functions are methods that belong to some class or instance of a class. This make things a little more complicated because now you cannot simply create a function called myOrder and pass it into a method because myOrder has to be a method of some class or object. You might think of defining something like: class Compare{ public myOrder(){};} class Compare{ public myOrder(){}; } and now you could try to pass and instance of the object: Compare myCompare=new Compare();myData.Sort(myCompare.myOrder); Compare myCompare=new Compare(); myData.Sort(myCompare.myOrder); That is you are trying to pass myCompare.myOrder to the Sort method. This is one approach and it could be made to work but C# is strongly typed and every parameter in a method has to have a type. What type is Compare.myOrder? Its a method of a type and not a type in its own right and trying to make it a type will quickly become very messy. A different alternative is to demand that every time you want to pass a function you have to pass an object that has that function as a method. This is what Java does for example. In this case we don't have a problem with types: Compare myCompare=new Compare();myData.Sort(myCompare); However within the Sort method the myOrder function has to be called as myCompare.myOrder(); In other words to pass a function you have to pass an entire object which wraps the function. Of course the one advantage of this method is that in principle you could pass a whole set of functions in one go but this doesn't make up for the inconvenience of the approach when you are trying to pass just one function. As any Java programmer will tell you this approach is workable but very verbose. You have to define a class for every function you want to pass and create and instance to pass every time you want to pass it. It is the original method that Java used to implement event handling and over time new facilities such as anonymous classes and eventually lambdas have been added to the language to make this cumbersome method easier to use. The C# Approach The overall C# approach to allowing functions to be first class entities has developed over time as well. Initially delegates were used as a way of wrapping a function in an object in a way that is superficially similar to Java's approach but it also has some special advantages. Over time delegates acquired features that made them easier to use anonymous methods and finally lambdas. However it is important that you fully understand the delegate idea and how to use them. Let’s look at how it works and some of the more interesting ways that you can put it to work. Delegates and their relationship to events is covered in another chapter. What is initially confusing is that to create a delegate you first have to create a type and then create an instance of the type. That is, delegate is a user-defined reference type that encapsulates a method. Consider, for example, how to encapsulate the simple example method: public int hello(int param1){ MessageBox.Show( "Hello delegate World" +param1.ToString()); return(++param1);} All this does is to display the current value of param1 and then return and incremented value. First we need to define a delegate type that matches its signature – including, in this case, the return type: delegate int HelloType(int param); This delegate type defines the methods that it can encapsulate. Notice that it is only the functions signature and return type that specifies the delegate type. The HelloType delegate can wrap any function that has the specified signature and return.. Notice that the default use of the invoke method makes the delegate instance look like a function object. That is HelloInst is an object but HelloInst(2) looks like a function call. Of course there are a few slightly hidden details in these examples which are obvious to experts but often confuse the beginner. The first is that the method being wrapped is still a method belonging to some object and not a "disembodied" function. When the delegate is used to wrap the method the method has to be accessible from where ever the delegate is being declared. In the example above there is the implicit assumption that all of the code is defined within the same class. A more elaborate and complete example would create the hello method within a new class: public class Greetings { public int hello(int param1) { MessageBox.Show( "Hello delegate World" + param1.ToString()); return (++param1); } } and then within the class that wants to make use of the method: delegate int HelloType(int param);private void button1_Click(object sender, EventArgs e) { Greetings g = new Greetings(); HelloType HelloInst = new HelloType(g.hello); int i = HelloInst(2); } private void button1_Click(object sender, EventArgs e) { Greetings g = new Greetings(); HelloType HelloInst = new HelloType(g.hello); int i = HelloInst(2); } Another subtle point is that. To wrap a method belonging to an instance of a class you have to first define a new delegate type with a specific signature and return type, then you wrap the method in and instance of the delegate type. <ASIN:1449380344> <ASIN:1430225378><ASIN:0470447613> <ASIN:1933988924> <ASIN:3540921443>
https://i-programmer.info/programming/c/9319-deep-c-delegates.html
CC-MAIN-2021-17
refinedweb
1,122
60.85
Domain Class Querying Querying with dynamic methodsThere are several ways to query for domain class instances, one of them being via Grails dynamic methods. For more details see DomainClass Dynamic Methods: def results = Book.findByTitle("The Stand")results = Book .findByTitleLike("Harry Pot%") results = Book .findByReleaseDateBetween( firstDate, secondDate ) results = Book .findByReleaseDateGreaterThan( someDate ) results = Book .findByTitleLikeOrReleaseDateLessThan( "%Something%", someDate )// find by relationship results = Book.findAllByAuthor( Author.get(1) )// and to affect some sorting results = Book .findAllByAuthor( Author.get(1), [sort:'title',order:'asc'] ) Querying by exampleJust pass an example of the domain object you would like to find to the find() method. def b = Book.find( new Book(title:'The Shining') ) Querying with a criteria builderFor more advanced queries or querying across objects graphs you can use Criteria (for a full reference see the section on Builders): def c = Book.createCriteria() def results = c { like("author.name", "Stephen%") between("releaseDate", firstDate, secondDate ) } Querying with HQL queriesOtherwise, as Grails uses Hibernate internally you can use an HQL query: Or with positional parameters: def results = Book.findAll("from Book as b where b.title like 'Lord of the%'") Or with named parameters: def results = Book.findAll("from Book as b where b.title like ?", ["The Shi%"]) You cannot combine named parameters with positional parameters.Be careful with Groovy multi-line strings: if they contain any newlines, then the query will not work. So for example this: def results = Book.findAll("from Book as b where b.title like :search or b.author like :search", [search:"The Shi%"]) will not work. You have to use the line-continuation character: def results = Book.findAll(""" from Book as b, Author as a where b.author = a and a.surname = ?""", ['Smith']) Note The space after "Author as a" is important, otherwise the string would look like "...Author as awhere b.author...". def results = Book.findAll("""/ from Book as b, / Author as a / where b.author = a and a.surname = ?""", ['Smith']) Querying with Eager FetchingSometimes is desireable to query objects that relate to each other without lazy=true , so SQL is optimized and there is no n+1 SQL SELECT queries. To do that you can:1) Add LEFT JOIN FETCH to your HQL query2) You can also try with Criteria Builder, with something like this: 3) You can use hibernate .hbm files to mapp relations using lazy=false. import org.hibernate.FetchMode as FM // ......def c2 = Task.createCriteria() def tasks = c2.list{ eq("assignee.id", task.assignee.id) maxResults(new Integer(params.max)) firstResult(new Integer(params.offset)) fetchMode('assignee', FM.EAGER) fetchMode('project', FM.EAGER) order('priority', 'asc') } Site Login
http://grails.org/GORM+-+Querying
crawl-002
refinedweb
432
53.88
- GraphQL - Step 2 - Updating lib/auth.js - Step 3 - Updating lib/tasks.js - Step 4 - Wrap Up Step 1 - GraphQL So far we've handled communication with the Amplication backend by making HTTP requests. However, Amplication provides another way of interacting with the backend, GraphQL. GraphQL is a querying language that allows for readable commands with many benefits. If you want to know more about why GraphQL may be a better choice for your application I'd recommend reading this article by the Apollo team. If you're running the backend ( npm run start:backend) you can tinker with queries on the GraphQL Playground To make the GraphQL queries to the backend we'll use a library called @apollo/client. First, install @apollo/client as a dependency in the web subfolder: cd web npm install @apollo/client We'll want to configure our GraphQL client. Create the following file web/src/lib/apollo.js and at the top of the file import @apollo/client. Then paste in the following code: import { ApolloClient, createHttpLink, InMemoryCache } from "@apollo/client"; import { setContext } from "@apollo/client/link/context"; const apiUrl = " const jwtKey = "accessToken"; const = createHttpLink({ uri: apiUrl, }); const authLink = setContext((_, { headers }) => { const token = localStorage.getItem(jwtKey); return { headers: { ...headers, authorization: token ? `Bearer ${token}` : "", }, }; }); export const client = new ApolloClient({ link: authLink.concat( cache: new InMemoryCache(), }); As in Tutorial Step 4, the @apollo/client has been configured to take the user's JWT access token and assign it to the Authorization header of every request. We'll also want to include the functions that check if an access token exists and to save a new access token. export const isStoredJwt = () => Boolean(localStorage.getItem(jwtKey)); export const setStoredJwt = (accessToken) => localStorage.setItem(jwtKey, accessToken); Finally, we'll want to export the gql from the @apollo/client. This allows for GraphQL queries and mutations to be written. export { gql } from "@apollo/client"; Step 2 - Updating lib/auth.js Open up web/src/lib/auth.js and delete all the code in the file. At the top of the file, we'll import some of the functions we created in the web/src/lib/apollo.js file. import { gql, isStoredJwt, setStoredJwt, client } from "./apollo"; Firstly, add the new me function: const GET_ME = gql` query me { me { id } } `; export const me = async () => { return isStoredJwt() ? (await client.query({ query: GET_ME }).catch(() => null))?.data.me : null; }; You'll notice that the query for the user account is broken up into two parts: GET_ME and me. The first variable, GET_ME is where the query is written. One of the benefits of GraphQL is that we tell the backend what data we want. In this case, all we need is the id of a user, so that's all these query requests. me will actually run the query. Next, add the login function: const LOGIN = gql` mutation login($credentials: Credentials!) { login(credentials: $credentials) { accessToken } } `; export const login = async (username, password) => { const result = ( await client .mutate({ mutation: LOGIN, variables: { credentials: { username, password } }, }) .catch(() => null) )?.data.login; if (!result) { return alert("Could not login"); } setStoredJwt(result.accessToken); return me(); }; Now, instead of referring to this as a query, we'll call this function a mutation. Queries are used to read data, mutations are used to write data. Logging in and signing up are technically writing data, as a session is being created in the backend. LOGIN is a mutation that takes the username and password of a user as an object and returns only the accessToken from the request. login will execute the mutation like the HTTP implementation. Instead of sending the credentials in the BODY of an HTTP request, credentials (and other arguments in general) are passed in a variables object. The key values of variables map to the variable names in the mutation we write. So variables.credentials in client.mutate maps to $credentials in mutation login($credentials: Credentials!). Finally, add the const SIGNUP = gql` mutation signup($credentials: Credentials!) { signup(credentials: $credentials) { accessToken } } `; export const signup = async (username, password) => { const result = ( await client .mutate({ mutation: SIGNUP, variables: { credentials: { username, password } }, }) .catch(() => null) )?.data.signup; if (!result) { return alert("Could not sign up"); } setStoredJwt(result.accessToken); return me(); }; Step 3 - Updating lib/tasks.js We'll next need to update the tasks functions to use GraphQL. Open up web/src/lib/tasks.js and delete all the code in the file and replace it with the following: import { gql, client } from "./apollo"; const CREATE_TASK = gql` mutation createTask($data: TaskCreateInput!) { createTask(data: $data) { completed createdAt id text } } `; export const create = async (text, uid) => { const result = ( await client .mutate({ mutation: CREATE_TASK, variables: { data: { completed: false, text, uid: { id: uid }, }, }, }) .catch(() => null) )?.data.createTask; if (!result) { return alert("Could not create task"); } return result; }; const GET_TASKS = gql` query tasks($where: TaskWhereInput, $orderBy: [TaskOrderByInput!]) { tasks(where: $where, orderBy: $orderBy) { completed createdAt id text } } `; export const getAll = async (uid) => { const result = ( await client .query({ query: GET_TASKS, variables: { where: { uid: { id: uid } }, orderBy: { createdAt: "Asc" }, }, }) .catch(() => null) )?.data.tasks; if (!result) { alert("Could not get tasks"); return []; } return result; }; const UPDATE_TASK = gql` mutation updateTask($data: TaskUpdateInput!, $where: TaskWhereUniqueInput!) { updateTask(data: $data, where: $where) { completed createdAt id text } } `; export const update = async (task) => { const result = ( await client .mutate({ mutation: UPDATE_TASK, variables: { data: { completed: !task.completed, }, where: { id: task.id, }, }, }) .catch(() => null) )?.data.updateTask; if (!result) { return alert("Could not update task"); } return result; }; Step 4 - Wrap Up Run the application and play around! Users' tasks are now being saved to the Amplication backend with GraphQL queries and mutations rather than traditional HTTP requests. Congratulations developer. Take with you what you've learned and build something amazing. If you need help or want to share what you're up to then you should join our Discord. To view the changes for this step, visit here. Discussion (0)
https://dev.to/amplication/amplication-react-using-graphql-235a
CC-MAIN-2022-21
refinedweb
964
50.02
Hi, I have class hierchary as follows. City.Province.Country So basically City has a property of type Province called Province and Province has a property of type Country called Country. In addition, each have a property of type string called Name. Now when I drag the City object from the DataSource windows, it binds the label Name perfectly. When I run the project I see the City and Province Name fine but for the Country I see the namespace of the Country class. The Text DataBinding property is set to CityBindingSource - Province.Country.Name Is it because BindingSource only supports properties 1-leve deep? Any feedback will be appreciated. Regards, Choudhry
http://www.windowsdevelop.com/windows-forms-data-controls-databinding/bindingsource-deep-properties-45875.shtml
crawl-003
refinedweb
112
66.33
Differences Between TypeScript Type vs Interface TypeScript is an open-source scripting language used for application development. The typescript trans piles to JavaScript and it is referred to as a superset of JavaScript. TypeScript compiler named ‘tsc’ written in typescript only and it is compiled to JavaScript as well. TypeScript is mainly used for developing both client-side and server-side javascript applications. An interface can be extended by other interfaces. TypeScript also allows the interface to inherit from multiple interfaces. An interface can be inherited in two ways that are Single interface inheritance and multiple interface inheritance. An interface is part of typescript only an interface can’t be converted to JavaScript. Let us study much more about TypeScript vs Interface in detail: The typescript was designed and developed by Microsoft in the year 2012. TypeScript has three components that are language, the typescript compiler, typescript language service (TLS). The language referred to syntax, keywords and type annotations. A compiler converts the instructions into javascript which is written in typescript. TLS is used to support a common set of typical editor operations like statement completion, code formatting, etc. TypeScript extending its functionalities with other libraries like Node.js, D3.js, JQuery, etc. Typescript follows its own syntax as it declares the data type of variable next to a variable name. TypeScript has a lot of features like compile-time checking, type inference, type erasure, interfaces, enumerated type, generic, namespaces, tuple, await, classes, modules, optional or default parameters, anonymous functions. It is more scalable and we can build robust components. It is independent of platform, browser and operating system also. It presents the method for the developer to express variables, arrays, and properties in a non-standard javascript way. TypeScript is superior to other scripting languages comparatively like Dart. TypeScript does not require a specific environment setup for the execution. It can be run where JavaScript can run easily. The benefits of using the typescript are that it will compile the code and check the compilation errors. It also finds the syntax error and tells before the running of the script. TypeScript has a feature of optional static typing and types inference system through the typescript language service(TLS). The type of variable can be inferred by language service (TLS) based on its value if a type is not declared for the variable. TypeScript types have a different set of types and values supported by the language. Types will check the allocated values to variables before storing it or executed for the application. Type can be of three types: 1. Any type: – By using any data type, it means type checking for a variable cannot be done. 2. Built-in type: – the data types which are already there in a system like a number, string, boolean, void, null and undefined. Null means that the variable has been set to an object whose value is undefined. Undefined means that the variable has no value or object assigned to it. 3. User-defined data types: – the data types which are declared by the user like enums, classes, arrays, etc. One of the types is also referred to as Type Alias TypeScript interface refers to the syntax that a system needs to follow. It is a virtual structure that exists within the context of typescript. It is mainly used for type checking purposes. It is simply a structural contract that defines the properties of an object is having like name and its type. An interface also defines the methods and events. It contains the only declaration of the members. Interface members should be declared by the derived class. Head to Head Differences Between TypeScript Type and Interface (Infographics) Below is the top 6 differences between TypeScript Type and Interface: Key Differences between TypeScript Type and Interface Below are the lists of points, describe the key differences between TypeScript Type and Interface: - TypeScript Type declaration can introduce a name for any kind of type including primitive, union or intersection type. Interface declaration always introduced the named object type. - The syntax for Type can be written as ‘type ABC = {a: number; b: number;}’. The syntax for interface can be written as ‘interface ABC = {a: number; b: number;}’. - In TypeScript, type does not create a new name for instance. In TypeScript, an interface can create the new name that can be used everywhere. - Type does not have a functionality of extending. An interface can extend multiple interfaces and class as well. - Type is mainly used when a union or tuple type needs to be used. In typescript, sometimes developers cannot express some of the shapes with an interface. TypeScript Type and Interface Comparison Table Below is the topmost comparison between TypeScript Type and Interface. Conclusion TypeScript type vs interface are the important concepts of Typescript. Type Aliases are sometimes similar to interfaces. Type Alias a primitive is not terribly useful, though it can be used for documentation. It can be generic like interfaces, where we can just add parameters and use them on the right side of a declaration. An interface has a feature of optional properties. These interfaces can be written similar to other interfaces. The optional property can be denoted with the symbol ‘?’ at the end of the name of the property while declaring it. The advantage of using this property is these can be available properties. It also prevents the use of properties which are not part of an interface. It has other features like use of read-only properties, excess property checks, function types, index types, class types and hybrid types Typescript is popular nowadays among the Angular developers and it is being widely used for developing the applications. The above is the difference explained between the typescript type vs interface that helps you to know about their basic things. Each of them having its own usage to work over the other. Recommended Articles This has a been a guide to the top differences between TypeScript Type vs Interface. Here we have discussed TypeScript Type vs Interface head to head comparison, key difference along with infographics and comparison table. You may also have a look at the following articles –
https://www.educba.com/typescript-type-vs-interface/
CC-MAIN-2021-04
refinedweb
1,027
56.55
Access ORC files from Spark Use the following steps to access ORC files from Apache Spark. To start using ORC, you can define a SparkSession instance: import org.apache.spark.sql.SparkSession val spark = SparkSession.builder().getOrCreate() import spark.implicits._ The following example uses data structures to demonstrate working with complex types. The Person struct data type has a name, an age, and a sequence of contacts, which are themselves defined by names and phone numbers. - Define Contactand Persondata structures: case class Contact(name: String, phone: String) case class Person(name: String, age: Int, contacts: Seq[Contact]) - Create 100 Personrecords: val records = (1 to 100).map { i =>; Person(s"name_$i", i, (0 to 1).map { m => Contact(s"contact_$m", s"phone_$m") }) }In the physical file, these records are saved in columnar format. When accessing ORC files through the DataFrame API, you see rows. - To write person records as ORC files to a directory named “people”, you can use the following command: records.toDF().write.format("orc").save("people") - Read the objects back: val people = sqlContext.read.format("orc").load("people.json") - For reuse in future operations, register the new "people" directory as temporary table “people”: people.createOrReplaceTempView("people") - After you register the temporary table “people”, you can query columns from the underlying table: sqlContext.sql("SELECT name FROM people WHERE age < 15").count() In this example the physical table scan loads only columns name and age at runtime, without reading the contacts column from the file system. This improves read performance. You can also use Spark DataFrameReader and DataFrameWriter methods to access ORC files.
https://docs.hortonworks.com/HDPDocuments/HDP3/HDP-3.0.1/developing-spark-applications/content/accessing_orc_files_from_spark.html
CC-MAIN-2018-47
refinedweb
268
50.02
On Mon, 12 Dec 2011 21:50:00 +0300Stanislav Kinsbursky <skinsbursky@parallels.com> wrote:> This routine is required for SUNRPC sysctl's, which are going to be allocated,> processed and destroyed per network namespace context.> IOW, new sysctl root will be registered on network namespace creation and> thus have to unregistered before network namespace destruction.> It's a bit suspicious that such a mature subsystem as sysctl newlyneeds its internals exported like this. Either a) the net namespaceswork is doing something which hasn't been done before or b) it is doingsomething wrong.So, please explain further so we can confirm that it is a) and not b).> --- a/kernel/sysctl.c> +++ b/kernel/sysctl.c> @@ -1701,6 +1701,13 @@ void register_sysctl_root(struct ctl_table_root *root)> spin_unlock(&sysctl_lock);> }> > +void unregister_sysctl_root(struct ctl_table_root *root)> +{> + spin_lock(&sysctl_lock);> + list_del(&root->root_list);> + spin_unlock(&sysctl_lock);> +}> +This requires the addition of a declaration to include/linux/sysctl.h.Once that is done and review is complete, I'd suggest that these twopatches be joined into a single patch, and that patch become part ofwhatever patch series it is which needs them.
http://lkml.org/lkml/2011/12/12/532
CC-MAIN-2015-35
refinedweb
184
51.68
Can we integrate SharePoint 2010 objects with standard Desktop application created using WPF or WinForm? SharePoint 2010 provides the Server Object model using which SharePoint WebSite Collection, WebSite, Lists and similar kind of objects can be accessed and programmed in client applications like WinForm, WPF etc. For this article, I am assuming that the readers know how to create a SharePoint List. I have created a simple website where Sales information is stored using the SalesInfo List as shown below: Note: Make sure you also read SharePoint 2010 Tutorials For Beginners and Intermediate Developers In this article, I am targeting the SharePoint 2010 List Object. Shown below is the object model for the List: The above object model clearly indicates that to program against the SharePoint ListItem (SPListItem) in the SharePoint, you need to iterate through the SharePoint SiteCollection (SPSite). The SPWeb represents the Web site which contains the SharePoint List (SPList). In the scenario below, I am using the SalesInfo SharePoint List for sales statistics of a product across various regions. This allows various sales persons to collect region wise sale data and update the List developed using the SharePoint technology. Creating a WPF Desktop client Application Step 1: Open VS2010 and create a WPF application called ’WPF_SPS_Server_Object’. Make sure that it targets .NET 3.5 frameworks since the SharePoint 2010 object model currently targets .NET 3.5. Step 2: In this project, add a reference to Microsoft.SharePoint.dll. You can find it on your machine on the following path: C:\Program Files\Common Files\Microsoft Shared\Web Server Extensions\14\ISAPI Step 3: Design the WPF application as below: Step 4: In the MainPage.xaml.cs, use the SharePoint assembly as shown below: using Microsoft.SharePoint; Step 5: In the Click event of the ‘Save List’ button, add the following code: The above code works with SalesInfo List in the ‘’ site collection. Please read comments in the code carefully. As you can see, the code shown above makes easy use of the SharePoint 2010 object model. Using this feature of SharePoint, a user can update the SharePoint application without physically using/logging into a browser. This feature is used by developers who are responsible to make new data entry on the SharePoint portal from their desktop applications. The entire source code of this article can be downloaded over here
http://www.dotnetcurry.com/showarticle.aspx?ID=758
CC-MAIN-2014-42
refinedweb
392
52.29
This action might not be possible to undo. Are you sure you want to continue? Passion of the Christ or Passion of Osiris: The Kongo Origins of the Jesus Myth By Asar Imhotep The MOCHA-Versity Institute of Philosophy and Research luntu/lumtu/muntu 1 A MOCHA-Versity Press Book Interior and Cover Art - Computer Graphics: Harold Johnson Copyright © 2011. 2 DEDICATION George Washington Carver (January 1864 – January 5, 1943), was an American scientist, botanist, educator, and inventor. This book is written in honor of one of the greatest Bakala Ngangas (African-American Masters), a true scientist and spiritualist; a true knower of Washil (Jesus) and how to access the Christ within; a spark in the bush. Ume Njalo: “May you stand forever” 3 TABLE OF CONTENTS Dedication................................................................................................................................................................................... 3 Acknowledgments ................................................................................................................................................................... 5 Introduction ............................................................................................................................................................................... 6 Linguistic Abbreviations....................................................................................................................................................... 9 Method ....................................................................................................................................................................................... 10 Understanding African Wisdom Traditions ............................................................................................................... 17 In The Name of Jesus? .......................................................................................................................................................... 27 Folk-Etymology............................................................................................................................................................. 30 Jesus/Osiris “lords of life”…and Water!................................................................................................................... 60 Bread of Life ............................................................................................................................................................................. 71 The Divine Shepherd ............................................................................................................................................................ 73 The Passion .............................................................................................................................................................................. 76 I am the way – The Cross and The crossroads .......................................................................................................... 77 Addendum....................................................................................................................................................................... 93 The Messiah ............................................................................................................................................................................. 98 The African schools of Life .............................................................................................................................................. 104 King of Kings ......................................................................................................................................................................... 109 Osiris, ESu, Jesus – The “way” ....................................................................................................................................... 115 Conclusion ............................................................................................................................................................................. 126 Bibliography ......................................................................................................................................................................... 148 4 ACKNOWLEDGMENTS When travelling on the superhighway of wisdom you meet many masters along the away; each providing you with tools to help better fulfill your purpose. I have met many such masters and anticipate more meetings of mutual enrichment. In this volume I would like to acknowledge two masters whom I have had the privilege of meeting. Oluwo Dr. Afolabi Epega: Educator, Chemist and 5th generation priest of Ifa. Ode Remo, Nigeria Shushukulu Dr. Mubabinge Bilolo: Egyptologist, Bantu-ologist, Linguist and Philosopher. Kinshasa, Democratic Republic of Congo I’ve only met Dr. Epega twice and each time very briefly. At the time I couldn’t appreciate who he was, but over the years I’ve grown to love and respect the work he has put in and the lessons he’s taught. They say you can tell a tree by the fruit it bears. I’ve gotten to know this tree by the fruits known to me as Babas Obafemi Fayemi (Obafemifayemi Institute for the Study of Ifa) and Ṣàngódaré Fagbemi (G.A.N Institute) who have introduced me to the wisdom of Ifa, through the teachings of Dr. Epega. Without your help and guidance, the connections in this book would not be possible. Ashe! Dr. Bilolo has been a tremendous help in guiding my understanding of Bantu thought, its languages and its history. I’ve always been attracted to the Bantu and never understood why until recent tests have confirmed my blood relationship avec la Famille de Cyena-Ntu, et surtout, à la Grande-Famille Éthiopienne, alias Famille des Langues Bantu via Musambik(e,a)/Musambuk(e,a). Meeting you confirmed for me that I was on the right path.1 Many thanks to the orishas, niombos, mimbos, nTrw, and ancestors who send me “hints” in dreams, and to my sun whose middle name is hrw for reason, I say: Ume Njalo! Ashe! 1 Although many guides have shaped this work, any errors are soley the result of the author. 5 INTRODUCTION Many publications have been written by various authors, from a wide range of disciplines, exploring possible connections between the central tenants of Christianity, Judaism and that of other religious systems in the geographical area surrounding ancient Palestine (Murdock 2009, Darkwah 2005, West 2009, Jochannan 1970, Oduyoye 1996; 1984, Issa and Faraji 2006, Baldick 1997, Pick 1913). Some authors assert parallels, while others argue for direct plagiarism; most publications arguing the Bible doing all of the “borrowing.” Movies like Zeitgeist is the source of much controversy over the internet as it alleges that Christianity is nothing more than a pagan solar cult which has borrowed stories from all over the Mediterranean. It doesn’t seem like the assault on Christianity will die down anytime soon as more and more evidence emerges that illuminate connections previously not considered, which places the originality of the Biblical narrative in doubt. This book was written in anticipation for a debate that took place on Saturday February 26, 2011 in San Antonio, TX between myself and a childhood friend, turned evangelist, TawfiQ Cottmon-El at the George Washington Carver Library in San Antonio, TX. This debate was a result of a series of “minor” debates on the popular social media forum Facebook, for which TawfiQ and myself are friends. This all started when TawfiQ, who is Christian, posted a video of a debate on a college campus between two noted scholars—one Christian and the other Muslim—on the topic “Is Jesus the son of God?” Another debate he posted was concerned with which book was the “true” holy book: the Quran or the Bible. The names of the presenters escape me now. As I was watching the debates I kept thinking to myself, “Why aren’t African scholars ever invited to these debates?” It is as if the hundreds of years of scholarship and the hundreds of scholars’ works dealing with the African origin of Judaism, Christianity and Islam has somehow disappeared, which allows Christians and Muslims to continue on as if there isn’t a body of knowledge that could upset their primary claims. This does not advance the field by any means. I noted to myself that once Africa is brought into the debate, it becomes a whole new conversation because what Christians and Muslims assume to be true as established fact, is put on its head once Africans begin questioning their major claims: especially on anthropological and linguistic grounds. From this point I asked that a debate be carried out and that Africa be brought to the table to join in the discussion. From there a mutual friend of ours, who goes by the name Ptah Hotep, set up the debate. It was initially supposed to involve myself, TawfiQ (a Christian) and a Muslim, but ended up 6 just being TawfiQ and I. TawfiQ suggested the topic be on whether Jesus was the son of God or not. I agreed to the debate topic. TawfiQ was to argue that Jesus was a historical figure and indeed was the Son of God as professed by modern Christians. I took the position that Jesus was a mythological character that was an adaptation of Osiris, and other deities from the continent of Africa. It was in preparing for this debate, within the last two weeks of the event, that I realized just how much information I would have to explain to the lay public for them to begin to understand the data that I would be presenting and the importance of the types of data that I would be using to make my argument. Coming to the realization that there would not be enough time to address everything that needed to be addressed in the short time allotted for the debate, I decided to create a document that would highlight my major points with supporting evidence. This work is the result of that effort.2 Parts of this document derive from original research that is going to be presented in an upcoming work titled Ogún, African Fire Philosophy and the Meaning of KMT. In researching for this work I discovered convincing parallels between the gods Osiris of Egypt and Eṣu of Yorubaland, Nigeria. Further investigations compelled me to cross compare both of these deities with Yeshua (Jesus) of the Hebrews of Jerusalem. My initial findings convinced me that there was a unique relationship between all three deities that could not be explained away as “chance coincidence.” Throughout the work on Ogún I make random references to these connections, but because the book isn’t focused on making these parallels, they only get mentioned in passing in support of a larger argument on the meaning of KMT. So I am glad I had the opportunity to address these parallels in a significant way in this work as it allows me to go deeper than I was able to in the work on Ogún. With that said, many who are familiar with my writings may see some repeats of information of past articles or books. The majority of the work, however, is unpublished data that has been synthesized with previous works that support my main premise: that Jesus is a myth based on an African myth. What do I hope people gain from this discussion? It is my hope that African-Americans and Africans in general, would have a greater understanding and appreciation for African philosophy, culture(s), spirituality, and its viability in our modern lives. As the 2004 Nobel Peace Laureate Wangari Maathai stated in her acceptance speech concerning culture: rediscover positive aspects of their culture. In accepting them, they would give themselves a sense of belonging, identity and self-confidence (Joubert and Alfred, 2001: 164) (emphasis mine) I apologize early if one finds many grammatical or spelling errors not caught by Microsoft Word. The short time constraint did not allow for me to hire a professional editor and have this document ready for the debate. I do, however, plan on expanding this work at a time where time is not an issue for which I can hire someone to review the work. So please charge it to my head and not my heart. If you find a mistake, please just email me and tell me what page can the mistake (s) be found and I’ll correct it. Hotep! 2 7 Many African-American Christians are averse to accepting the kinds of data that I present in this work because they do not find much value in who they are (culturally) or in the contributions their ancestors made historically. Racism in this culture has taught African-Americans to shun anything that is associated with themselves and to elevate ideas that derive from European sources. As a result, when faced with evidence that the religion they love with all of their heart is in fact an African “pagan” religion, they shut down and refuse to accept what has been presented to them. Maybe if someone of European descent tells them, then maybe they’ll listen (West 2009, Bernal 1987, 2005, Campbell-Dunn 2004, 2006, 2008, 2009a, 2009b, Bauval 2011). So much of the confusion surrounding this debate is based on Black people devaluing their collective and historical worth. The vast majority of Blacks in America, who are Christians or Muslims, do not respect their ancestral contributions, and in turn do not respect themselves. With this short work I seek to: o o o o o o o o o o Plant a seed that would inspire deeper study of one’s own heritage End the historical dark-ages of Biblical scholarship Foster a new era of understanding based on sound scholarship End the needless proselytizing which has wrought catastrophic disharmonies on the world stage in African communities Bring balance to the world Respect and acknowledge the ancestors who stories have never been told Inspire humanity to be lovers of wisdom Bring the restoration of nature back into the public discourse and into popular culture. To reorient our collective consciousness back to an earth-centered consciousness. Present jewels of wisdom that will allow the reader to fertilize their own seeds of greatness So I hope you find this work informative and edifying and I look forward to your feedback. Asar Imhotep MOCHA-Versity (Kala ba Nganga) 8 LINGUISTIC ABBREVIATIONS PB PWS PWN PNC PCS PAA PPAB Bantu BANTU “Bantu” A-A ES CS CN NS Proto-Bantu Proto-Western Sudanic Proto-Western Nigritic (Westermann) (Mukarvosky) Informal. No systematic reconstruction available Proto-Central Sudanic (Bender) Proto-Afro-Asiatic (Ehret, Diakonoff) Proto-Potou-Akanic-Bantu (Stewart) Proto-Bantu (Meeussen, Meinhof) Common Bantu (Guthrie) Bantu & Semi-Bantu (Johnston) Afro-Asiatic (Diakonoff, Ehret, Greenberg) Eastern-Sudanic (Greenberg) Central-Sudanic (Greenberg) Chari-Nile (Greenberg) Nilo-Saharan (Greenberg) [I have used Greenberg’s abbreviations (numbers & letters in brackets) to identify languages]. N-C Mande TogoR Polyglotta Niger-Congo B Banbara, D Dioula, M Malinke (Delafosse,Westermann) Togo Remnant (Heine) Koelle’s Polyglotta Africana 9 The Librarian states. These people are so attached to this film that they judge whether they can be friends with you based on if you love this film or not. If you don’t like this film. you are unevenly yoked and nothing but bad luck will surround your life if you do not accept this film as the one and only film that can change your life. The Librarian agrees with the last point made by the Bishop. Thelma Carpenter. The film follows the adventures of Dorothy. Befriended by a Scarecrow. They name their children after characters in this movie. she travels through the land to seek an audience with the mysterious "Wiz". make the likelihood of The Wiz being an original film highly improbable. Imagine a population of people so in love with this film that they decided to use the themes in this film as their spiritual gospel. Therefore. and Richard Pryor. The Bishop then goes on about how the film made him feel and the positive impact that it has had on his life and the lives of his family and congregation members. The Bishop agrees and now it is time for the debate. similarity doesn’t necessarily mean causality. the Librarian invites the Bishop to a scholarly debate where each side can present their evidences that support their respective positions. The Librarian then asks the audience to find an explanatory model to explain how there was no borrowing on the 10 . Ted Ross. who they say has the power to take her home. The Wizard of Oz is set in Kansas and the cast is all Anglo-American. Now imagine that some old librarian stumbles upon an old copy of a film called The Wizard of Oz with essentially the same theme that was created in 1939. It is now the Librarians turn. that this isn’t a debate about mere similarities. This movie to them is so original and unique that all other films are measured in terms of this film. and a Cowardly Lion. even years after the film was made. a Tin Man. however. these are two different stories and any claims of borrowing are not grounded in truth: any similarities can be chalked up to coincidence. Firstly. He simply recites reviews of the film made months. Lena Horne. New York schoolteacher who finds herself magically transported to the Land of Oz. Frustrated with the way this conversation is going.METHOD In 1978 a musical film came out called The Wiz starring Diana Ross. The main character of The Wiz is a school teacher. Michael Jackson. The Wiz takes place in Harlem and the cast is primarily African-American. The old librarian presents this film to the chief Bishop of the Church of the Divine Wiz and it is immediately rejected on grounds that they believe in the originality of The Wiz and it is impossible for The Wiz to have been borrowed or inspired by any earlier films. The Bishop goes first and doesn’t present any evidence that The Wiz was the true and original story. To the Bishop. Mabel King. Their whole lives are based on this movie. Theresa Merritt. while the main character of The Wizard of Oz was sent to a magical land by way of a tornado. a shy Harlem. which resembles a fantasy version of New York City. The Bishop claims that just because there are similarities in both films doesn’t mean that one borrowed from the other. The main character of The Wiz is sent to a magical land by way of a snow storm. but a sequence of exact features that when added up together. This was an excellent film and one of my alltime favorites (the first being Berry Gordy’s The Last Dragon). while the main character of the Wizard of Oz is a farm girl. Nipsey Russell. same characters and the same title prior to the release of The Wiz and still maintain that The Wiz is the original creative work? The Bishop then replies that the Devil (the Wicked Witch of the West) has the power of prophecy and prophesied to the creators of The Wizard of Oz and inspired them to create the film earlier in an attempt to thwart the plans of God in The Wiz. This is unnecessary as the argument is that. The Librarian goes on to point out that both films have a yellow brick road. that both characters are trying to find their way home from Oz and that Dorothy meets up with a tin man. It is the belief of those who adhere to the Abrahamic religions that unless one can prove exact correspondences on every point imaginable between two or more religions. modern Christians behave like the Bishop with The Wiz when it comes to the evidences presented to them that Christianity. but an urban adaptation of The Wizard of Oz. combined with other features that prove that The Wiz is not an original film. like The Wiz. but that all of the characters in the 1939 film are present in The Wiz which came out in 1978. or that there is a slight variation in the characteristics in the main characters of the narratives. that both of the main characters wear ruby red slippers. Christianity is an adaptation of earlier myths. the Wicked Witch of the West The Librarian points out that a few of the names are slightly different. So the Librarian challenges the Bishop to explain how do you find two previous mediums with the same themes. Christians try to deny the connections between the Biblical stories and those found in other ancient myths by claiming that the stories don’t exactly play out the same. has its roots in other previous spiritual traditions in the geo-political area of the “Near East” in ancient times. At this point the Librarian raises his head and hands in disgust and walks off stage to go get something to drink. then one cannot claim borrowing from one another. This is 78 years prior to the creation of The Wiz.side of The Wiz from the Wizard of Oz film when the names of the primary characters are the exact same in both films: Main character names in both films The Wizard of Oz Dorothy Gale Hunk/ Scarecrow Glinda the Good Witch of the North Zeke / Cowardly Lion Hickory / Tin Man Aunt Em Uncle Henry Professor Marvel / Doorman / Cabbie / Guard / The Wizard Miss Almira Gulch / The Wicked Witch of the West The Wiz Dorothy Gale Scarecrow Glinda. The Point Although this is a fictional story above. that both characters get lost in a place named Oz. The way the narrative plays out doesn’t have to be the same. 11 . scarecrow and lion. The Librarian also points out that The Wizard of Oz was based on an even older book titled The Wonderful Wizard of Oz written in 1900. the Good Witch of the South Cowardly Lion Tin Man Aunt Em Uncle Henry The Wizard/Herman Smith Evillene. as we know it. What is the likelihood that in both films the main character would be named Dorothy Gale? The Librarian points out that although there are major differences in the film. Just because one film is set in Kansas and the other in Harlem doesn’t diminish the other exact thematic matches described in this debate. it is the sequence of the main characters. 25 "they exchanged the truth for a lie. What the author has just emphasized is the importance of chronology.".1:20-25 tells us that man from the beginning knew God.. Several tools are available to us that help us to better analyze and interpret history. he asks: "Where did the pagans get a concept of three? Why not two or four ? Where did they get the idea of a God in heaven anyway? What about their belief in a virgin and a son. Is the Trinity Pagan?. In this website the author makes the following statements and claims concerning the Trinity." So in other words. then a Christian has to find an explanatory model to explain how this could be since Christians claim that Christ only came to earth once (at least as Jesus) to die and resurrect for the sins of mankind.What’s important for comparative studies is the number of major themes that are present in two or more stories that make the case for borrowing. all of the ancient traditions have a trinity concept because the god of the Bible revealed himself to them as such at an earlier remote time. Let’s just say for the sake of argument that the author’s contention is true. The tool used 12 . When TawfiQ and I were first debating online. they must come to the God of Israel (and Christianity in general) to get the full understanding. then the Christian argument is null and void and the mythology of Jesus is established. The question then becomes one of chronology: which story came first and which story influenced the other. he cannot make the same argument for Jesus the Christ. This means that ANY attestation of Jesus.letusreason. if they want to get the “true” understanding of the trinity. While he can make this argument for the Trinity. as Jesus (a Christ). and worshipped and served the creature rather than the creator.htm). geology and biology (genetics). In this exchange TawfiQ sends me to a website in defense of one of his positions (See. philology. it can no longer be argued that both stories developed independently of each other: especially if it has been proven that there was contact between the two cultures." When mankind fell into darkness of sin. they are no longer capable of understanding the true nature of the trinity and therefore. God decides to withdraw from them and then chooses to go to the Hebrew people to “represent” himself unto them because he knew that they would “get it right. If we find Jesus going through similar adventures in earlier traditions. Jesus is believed to have existed among the Hebrews at a specific time period in history. Once a deified Jesus is found in earlier traditions. they still retained some elements of the truth but distorted its meaning and it became lost. “BS.” The other ancient people have a trinity. Let’s just give him that..” I actually thought this to be a very good thing that Tawfiq presented me with. Besides the immediate reaction of. This case for borrowing is strengthened when it can be demonstrated that these major themes are carried out by characters with the exact same names like the cast of characters in both The Wizard of Oz and The Wiz. That is why this work is focused on finding JESUS at an earlier time and not similarities between similar gods in other traditions with different names. These tools include archeology. Because these people’s heart was wicked. Our Roman calendar is based off of the alleged birth of Christ which ranges from 4 BCE to 6 CE. 'and their foolish hearts were darkened" vs. among another people. our subject was on the validity of the Trinity Doctrine in certain camps of modern Christianity. at an earlier time pokes holes in the whole Christian argument. anthropology. After identifying the main characters in both traditions as being the same. where did that originate from? Rom. but because of their evil ways. The historical. In a featured article in the Biblical Archaeology Review (5-6/00) entitled. In this work we also seek to demonstrate how African myths work so that we are able to better analyze the Christian narrative using the tools the ancient Africans used to create their stories in the first place. by scientists and the Biblical narrative itself. literary and archaeological evidence of the influence of Egypt on the Levant is abundant and includes the presence of Osiris in Israel. The development of the ancient Egyptian society is at least 2500 years older than the “Hebrew” people and this gives the Egyptians plenty of time to develop religious narratives rooted in their cultural world-view and enough time for the Hebrews to find inspiration by the Egyptian imagination. According to the Hebrews themselves. when it was under Egyptian domination. one forces them to explain how Africans got a hold of Jesus before the Hebrews and yet claim that the Hebrew narrative is the original story. in name and in major characteristics. . It is my contention that Jesus is a composite of many features that is embedded in his name that belongs to a theme of ideas encapsulated in an ancient root system for which only historical comparative linguistics can reveal. in Africa prior to 4BC or 6AD. in the Gaza Strip. then Christianity has a major problem because their argument is that Jesus was a real person that existed at this time period. they developed as a people in Egypt. “How many coincidences are we allowed before we can say that the coincidences are no longer coincidences. It seems that this Egyptian deity was especially popular in Canaan. With evidence like this. If we find Jesus. but solid correspondences (facts)?” This question should be kept in mind throughout the duration of this discussion. which would mean that they would be very familiar with the conceptual ideas of the Ancient Egyptians. At the time the Israelites allegedly left for Canaan. How can Jesus have come to the Hebrews.primarily in this book is that of comparative historical linguistics. in addition to many of the tools mentioned above. What one has to do is make the Christian explain how we find the same narrative in Africa. no Christian can claim that the ancient Hebrews did not have an opportunity to be introduced to the deep tenants of Egypt’s Osiris: especially when temples of Osiris existed in Canaan—the land the Hebrews bragged and boasted about conquering and dwelling in (the promised land). and we have records of him in ancient Egypt saving the Egyptians 2500 years prior? This is not something minor and Christians know this and is why they try to avoid the chronological issue altogether and focus on the impact Jesus has had on their lives. It is an undisputed fact. that the ancient Egyptian society existed before the Hebrew people. 13 . That is trying to play the game by their rules. it was under the rule of the Egyptians. its language and the political state of Israel. This puts a major cloud of doubt on the whole Exodus story. and the earth for the first time in order to “save” it. Why would you escape Egypt only to go to Egypt? What we are essentially asking the Christian at this point is. thousands of years before the Hebrew story? In other words. we do not have to get into a circular debate on the evidence of his existences in Palestine. This is why linguistics is important because we can establish similarities on a much stronger footing. When discussing the originality and the authenticity of Jesus the Christ among Christians. on the basis of recurring patterns by the analyst…Ultimately. Continuing on pg. the more powerful it becomes. but through the power of custom. Furthermore.Jesus the Christ represents a theme. The name gives the root metaphor permanence and therefore it can do its job many times over. or reformulating an idea can be tried out. The greater its ability to incorporate and adapt to new experiences. When a psychiatrist says his patient is suffering from Oedipus complex he can name and summarize what otherwise might take pages to explain. and he can use the label repeatedly. Haitians were free to adjust or augment its content…As such. not a single cluster of meanings…The unities that exist in the diverse body of Ogun representations—as revealed by redundant themes found in many societies of the two hemispheres—are achieved not with the help of a supporting bureaucratic. a generalization about the phenomenon under consideration already has been made. interpretations that can be verified. When we examine the cognate terms for the name Ogún. is a “root metaphor. once a root metaphor is named it becomes a protected category within which many ways of replicating. Barnes conceptually understands the importance of roots in comparative religious discourse. Jesus. A first step is to find redundance. Given the problems with drawing boundaries. how can we. they condense into one label a complex historical essay on the uses and abuses of Power. this generalization lends itself to interpretations. thanks to redundance. is concentrated in what is now more broadly known as the symbolic school. we begin to notice a common set of themes which can be demonstrated linguistically by examining the g-n root that makes up the name Ogún. codified apparatus. Like those who work with texts. It is my contention that many of these root metaphors are “rooted” in linguistic consonant root systems. when Haitian devotees call a despotic leader Ogún Panama. 1997: 1415). By the same token. the key to a middle-level analysis has come to rest primarily in their use of methods. delineate domains? [the themes]. that can be confirmed either by the action or. We discover what the theme is by examining redundancy of the shared themes in the myth. recommended above. rituals. metaphors.” I argue the same for Osiris (Wsr) of the ancient Egyptian tradition. Needless to say. Ogún Panama was first applied to one person. humanist and social scientist adherents of this school try to do two things: interpret cultural representations found in oral traditions. that elicit redundancies in cultural meanings and. translate their findings into terms that can be appreciated by outsiders…For symbolists. 20 she notes that: Ogun is a metaphor (see Pepper 1966:91-92). later other despotic leaders were given the same label. The study of custom in unwritten traditions. Although not speaking linguistically. restating. while conventionally the preserve of anthropologists. (Barnes. A root metaphor names the things that are likened to one another. then. after a real figure. I argue. 14 . Barnes (1997) explains how this is done when examining the Yoruba deity Ogún who also is a composite of ideas based on a linguistic root. analytically. All mythological characters that I am aware of can be explained using the “root metaphor” concept. By naming the metaphor. we are looking for redundant clusters of meanings. Sandra T. symbols and artifacts in ways that are faithful to the actors and. sexual prowess and urbanization. a “root metaphor” that is based on lexical terms that share the same root theme because they share the same root consonant system: g-n. to stay in place” Qemah “flour” gun “to pound” (pulverize) gun yan “to pound iyan” agunmu “medicinal herbs” (pound into a powder) agan “barren (woman)” Ogún “god of iron” Aqama “sterile” Qayn “smith” Qayin “Cain” The Hebrew Qayin “Cain” is the Yoruba Ogún “god of iron and farming.” We have two Qayins in Genesis: Qayin “farmer” and Tuwbal Qayin “blacksmith. is going to be very important in this study for those readers not familiar with linguistic comparative studies.” Ogún takes on all of these characteristics. Although the initial consonant is different for the same word in Yoruba and Hebrew respectively. farming. war. but Ogún represents a concept.”Ogún is the god of “fire. iron. By cross comparing related languages. “it is long” O gun “it is erect” O gun „yan “she is protruding at the breasts. a theme. we consider them cognates because the sound matches correspond on a regular. Her breasts have become erect.” Ogún “war” Gun le “to settle” a-da-gun-odo “stagnant waters” Arabic Qaama “he rose” Qomah “height” standing place Maqama “combat” Aqama “to settle. This type of analysis. we can establish that the Yoruba /g/ in initial position corresponds to Semitic /q/. not because this spirit necessarily is all of these things. 15 . stand up” Yoruba Gun (oke) “climb” (hill) gun (akaba) “climb” (a ladder) gun (esin) “ride” (a horse) O gun “he is tall” i. Further explorations within the Yoruba language help to demonstrate sound shifts within the Yoruba language itself which gives us additional information for where to look for correspondences in other related languages. the analyzing of sound shifts. systematic basis.Hebrew Qum “rise up.e. hunting” /k/ Akin okun akoni okàn kàn kúná kun okan èkun ikun ikín ekun kìnníún gún “to pound” ogún “twenty” (cf. they also share the same features or “themes. war. Hausa gama) gún “finished product”. s-l) that embodies an “associative field” of meaning. This makes it triply difficult for Christians to argue originality when the same “root metaphor” (collection of themes personified as a deity) Jesus can be found in 3 areas of Africa.g <> k Alternation in Yoruba /g/ Ogún “God of iron. The first is among the Yoruba of Nigeria. This is important because we can use this linguistic tool to establish that Jesus and Osiris are in-fact the same deity. s-r. the probability of chance correspondences is seriously diminished to virtually zero. When we line up all of the features and common themes associated with Osiris and Jesus. 16 . limits” “moisture in the nostrils” “stalk. These themes are captured within a set of linguistical roots (-s-. we must first introduce and analyze African wisdom traditions: the school that is responsible for the way myths are told in Africa. Nothing will make sense until the reader understands African traditional education from the standpoint of the Africans who live. one. We now begin our primary discussion. edge. and whose personified themes existed in the area prior to colonization and the introduction of Christianity by Europeans (or Islam by Muslims). because they share the same linguistic root. brave. extremes” òógun “perspiration” (moisture) ègún “thorns” ológìnní “cat” Meaning “bravery. Before we get into the meat of this debate. reed” “tiger” (cat family) “lion” (cat family) Throughout this work I will demonstrate that all of the epithets and characteristics of Jesus is so because this mythological figure was created by the same process for which the “root metaphor” of Ogún was created. and the other is with the Baluba-Bantu of the Democratic Republic of Congo. eat and breathe it on a daily basis. To ensure that coincidence does not play prominently in our investigation. thousands of miles apart. valor” “strength” “brave person. grade1 igun “angle. Having this foundation will provide the reader with the necessary tools to properly analyze the Christian narrative and see it with fresh eyes. I have chosen to examine two other pre-Western societies that also have the Osiris “theme” in Africa. Because they share the same root. hero” “(lion) heart” – heart “to hammer” to be powder – smooth “to be full” “1” – integer “ends.” I argue that this is the case because ancient sages deliberately wanted to highlight some concepts for the edification of the ancient people and chose to encapsulate all of these themes under one “root metaphor”: in this case Osiris and Jesus. UNDERSTANDING AFRICAN WISDOM TRADITIONS Since the dawn of time. So our true nature is spiritual. Sages of African communities of memory develop worked-out approaches and solutions that display wisdom traditions as coherent gestalts. the search for wisdom has been the most important undertaking for human beings who accumulate it and pass it from generation to generation. every person is an incarnation. mountains. rivers and still water. In the mind of indigenous people across the world. that is. Grandfathers and grandmothers. ancestors assist human beings in making life more beautiful on earth.” What concerns us primarily in this discourse is the pre-Western pedagogical methods for transmitting wisdom from one generation to the next in communities of memory. To understand African communities of memory is to understand the role of ancestors and their visionary leadership. They are the keepers of the very wisdom the people need to live by. Centers of wisdom produce values and services that determine the basic priorities of what challenges to deal with and what social ends to serve in the communities of memory. a spirit who has taken on a body. Could the name Bashi be the name the Ancient Egyptians used to describe the dwarf god Bes? The Egyptians used to specifically seek out these short statured people who in ancient times used to live from the Congo to Ethiopia. he underscores the importance and the role of the ancestors according to his wisdom tradition: For the Dagara. In the Democratic Republic of Congo.” Three main themes emerge when we explore the written and oral literature of African wisdom traditions that illuminate their purpose and function in pre-Western communities of memory. 3 17 . Peter Vail (1995: 64) provides a good definition of vision that can be applied to African communities of memory: “The ancestral vision is a vivid expression that describes why a particular community exists and what types of human beings it intends to deliver. They are: 1. the Bashi3 people have a proverb: “Wisdom is the only eternal thing. He defines it as. Ancestors are the architects of communities of memory. A community of memory is an organization whose encompassing vision is the search for wisdom as defined by its ancestors. Kykosa Kajangu (2005: 88) provides a working definition of the study of wisdom traditions. usually an ancestor that somebody already knows. They are the advocates of humanity in the womb of creation. 3.” Dr. The life energy of ancestors who have not yet been reborn is expressed in the life of nature in trees. Ancestors establish moral understandings and spiritual commitments that bind communities together. Sacred arts embody and express wisdom and illustrate the ways challenges are posed and solved within communities of memory. “The exploration of pre-Western modes of thought and being that are embodied and displayed by sacred arts and that make it possible for sages of communities of memory not only to know life but also to respond of its challenges. who has important tasks to do here. A birth is therefore the arrival of someone. According to Malidoma Somé of Burkina Faso. The ancestors are the real school of the living. 2. are as close to an expression of ancestral energy and wisdom as the tribe can get (Somé 1994: 20). This world is where one comes to carry out specific projects. therefore. in his work Of Water and the Spirit. in his work Indaba My Children (1964) supplies us with an authentic basis for communion with the ancestors. 1964: 573). lineage. 2005: 94). reinforces the relationship of ancestors and the soil among the Dagara when he states: “As we walk the earth. In this perspective. called Bwami. The verb cudima has spawned three other terms that weave interrelated concepts together to form an associative field of ideas. Credo Mutwa. function or place within an African community of memory of these institutions and I seek to analyze these organizations here as they are critical for our discourse throughout this work. All three concepts are united by the idea of adding life to life through productive activities. forcing the people and their activities to remain a secret. These endeavors are guided by the ancestors whose spirits reside in the very environment we seek to understand and have experiences. a Zulu shaman. This belief that a man lives solely to serve his ancestors is one of the most deep-rooted beliefs in the whole of Africa. We continue our exploration of the role and importance of ancestors by examining the wisdom tradition of the Baluba people of the Congo. It has the structure and some of the functions of a voluntary association. Only the hard work of the brightest talents can move mountains and bring wisdom. Access to and advancement in bwami are conditioned by a number of factors: character. in his work The Healing Wisdom of Africa. The phrase “centers of wisdom” is coined by Kajangu in replacement of the inaccurate label “secret society. kinship 18 . but it also maintains and reinforces kinship. Malidoma Somé. This label does not convey the real role. and budimu (wisdom). through the medium of an agricultural term cudima.” Secret societies are labeled as such only in recent times because the once open education systems across Africa had to go underground during the colonial period. which varies with context and usage. expressing the idea of labor. They provide guidelines for successful living. He goes on to note: Bwami is many things in one. which I feel essentially applies to all African wisdom centers based on my comparative studies. Communities thrive when the power holders are the invisible exemplars whose life-stories embody and display the best approaches and solutions to meet the many challenges of life. It is with this cultural understanding that we may be able to penetrate the true spirit of the stories as expressed in the Biblical myths. In his work Lega Culture (1973) he provides a synopsis of the Lega tradition. It is this belief that fuels the practice of deifying ancestors in Africa who have made great contributions to society. The verb cudima is composed of three elements: the infinite particle cu. and the –a suffixal ending. and tribal unity is based on this. African Centers of Wisdom Kajangu (2005) provides an exemplary discourse on the function and role of African centers of wisdom for which I draw inspiration in this section. 1998: 169). how do African centers of wisdom enable the muntu (human being) to know life and stem the tide of its challenges? Daniel Biebuyck looked for wisdom among the Lega people of the Democratic Republic of Congo. and clan bonds. The tribe as whole must keep the spirits of its founders alive—every tribe in Africa believes this (Mutwa.Ancestors are the pathways between the knowledge of this world and the next. all relating to agriculture: budimi (field). mudimu (work). The question we ask here is. we are warmed by the heat of the ancestors coming from the underworld below us” (Somé. there is the greatest value in the soil where ancestors are buried (Kajangu. the radical dim. divination 19 . These centers constitute schools for lifelong learning. dances. Human beings are shelving instruments that hold data. Therefore. there would be no wisdom traditions in which the brightest minds of the community can provide enlightened leadership to guide others in knowing life and in stemming the tide of its challenges. Their lives embody the teachings of the wisdom centers. Centers of wisdom teach that in order to be wise. Centers of wisdom provide strategies to assist human beings in their search for perfection. invests and reinvests it. It is the youth’s job to expand the collective pool of wisdom. uses and explains thousands of pieces of sculpture. Our jobs. To cultivate this stretching of the imagination. Without wisdom centers. The initiations aim at moral perfection. distributes and redistributes it. the principles of which are elaborately explained in proverbs. wealth. It is also an arts club. human beings construct wisdom through working together in special organizations. its dances and musical styles. 2. Wisdom centers are created to organize this accumulated human intellectual wealth (what they call in the Kongo Muyudukwa) for proper usage in the communities of memory. The leadership in the community is always held by wise elders. The information needed to navigate this plane of existence exists outside of oneself. an elder has a proven track record of success in taking direct responsibility for the development of the youth in their community of memory. They have developed countless strategies or teachable viewpoints to take people to places where they have never dared to go. royal staffs. and initiation. Rather. It is a school of art because it creates. These sages have learned to the highest degree the secrets for knowing life and stemming the tide of its challenges. thinks and ponders before moving horizontally to meet the challenges of the diurnal world. Memory boards. the sacred arts were invented to train the mind to see inbetween the hidden layers of reality. one must be guided and learn to guide others. These are individuals who are masters at building symbolic worlds which aid in the management of the destinies of African wisdom traditions. In essence. This is important to understand as the job of an elder is to help younger generations to discover new more satisfying dimensions for being human. wisdom centers act like spider webs which connect and give meaning to all the dimensions of existence in the community of memory through the sacred arts. ancestor figures. We then ask ourselves. for it enjoys and patronizes the fine arts. and provides economic incentives. how is the teaching and learning of wisdom to be orchestrated in these unique communities? According to Kajangu. dramatic performances. 3. Bwami is a religion without gods. It is like a big corporation that produces wealth. produces. choreographies. the human being stands vertically. as elders in training. pretending to have a power of its own and to master the secrets of making life good (Biebuyck 1973: 6667). Bwami has developed its own literary arts. are to travel to the shores of eternity and record what we witness and find ways to translate into experience that which we witnessed in the material world. In other words. The wisdom tradition among the Luba provides a perfect example of how the sacred arts mold its community of memory.support. and architectural styles. For African people wisdom does not fall from the sky. It is this reservoir of knowledge that enables elders to be effective teachers. How is enlightened leadership promoted? Initiation is the method par excellence in Africa for creating enlightened leadership. who Kajangu calls centers of wisdom. and objects. wisdom centers fulfill three major functions that provide a framework for students of wisdom to know life and to stem the tide of its challenges: 1. This means that centers of wisdom are schools for enlightened leadership. elders are living examples on how to reach the summit of achievement in any given society. Lastly. This ‘Luba memory’ is going to be important for us throughout this discussion as the Luba community still practices the ancient wisdom which can be seen in Ancient Egypt and in Nigeria which are places I use to cross compare and verify correspondences between Egyptian and Hebrew: the focus of this very work. wooden memory boards. Sages in wisdom centers attempt to unravel the mysteries of the universe and the place of human beings in the grander scheme of things. the sacred arts are used as mnemonic devices to transmit knowledge. the centers of wisdom work as esoteric schools which initiates penetrate to the depths of knowledge. who could recite genealogies. figures and staffs: and scepters. still holding on to lexical items and noun class variations from Proto-Bantu. The State of Perfection The most important function of African wisdom centers is the mission of producing ideal human beings who have realized the state of perfection and who live in harmony with the central purpose of creation (Kajangu. It is through teaching and learning that centers of wisdom enable human beings to know life and to respond to its challenges. Roberts & Roberts (1996: 33) note. These centers of wisdom act as laboratories where sages are always busy experimenting with sacred arts. and other Luba mnemonic devices dating from the eighteenth to the twentieth century demonstrate the importance of visual arts and related expressive culture to the formation. Among the Luba of Nkuembe and Buadi. thrones. The learning process was not designed for a short period of time. they are scientists. all of the cultural motifs are used to teach history. It is a life-long quest in search of wisdom and perfection that never exhaust itself. The Tshiluba language. is one of the most conservative Bantu languages. In regards to the Luba memory. According to oral sources in the Nile Valley. they also possess a proliferation of visual forms to encode and stimulate mnemonic processes. They are not objects of worship as many Christian historians have articulated. We now move on to discuss centers of wisdom as institutes for life-long learning. axes. In essence. 20 . and other objects incorporating iron and/or copper. spears. They travelled with kings…and spread propaganda (Roberts & Roberts 1996: 37). If Luba have a rich vocabulary to express concepts of memory.” These sages only excel with the help of other sages. They also explore the physical world in which human beings live. for example. 2005: 103).instruments. They go on further to state that “Mbudye historians were rigorously trained ‘men of memory’. king lists and all of the episodes in the founding charter of kingship. In other words. It is these organizations which the custodians of African wisdom 4 The root of this term will become very important later on in this discourse. development.4 which means “the most accomplished sages. In other words. the ancient Egyptian traditional education was a minimum of 40 years. Those enlightened students go on to obtain the title bantu basunguluke. It is because of these methods in African societies that I feel the oral tradition has been an excellent craft for which ancient knowledge is passed on. and remembrance of Luba kingship and political relations (Roberts 1996: 17). It is a form of peer pressure that encourages success through mutual enrichment. adzes. These include beaded necklaces and headdresses. They utilize all that they have learned through these explorations to look for ways and means that enable the muntu to achieve perfection. All aspects of wisdom accumulation are embedded in the culture and its motifs. but would use the powerful voice of his mind to speak to both beasts and fellow human beings. Culture. Hunt Originally published in French as Aspects de la civilisation africaine: personne. Malidoma Somé claims that one must be initiated. There would be nothing hidden from this man [or woman] who would be using the forces latent in our minds to do even the most seemingly impossible things. we hope. This is the fundamental difference between what one observes in Africa versus what one witnesses in the Bible as the Bible appears to be anti-science and learning. Such a [human being]. When a child grows into an adolescent. And. 1972. my son. and this is a Amadou Hampâté Bâ. instead of using spears and shields. It is here where the minds of initiates are enriched and their heart finds solace. in Africa. are acquired in these centers of wisdom which aim at perfecting the human being. African wisdom centers develop sages through a process of initiation. and Thought of Traditional Africa. It is my belief that perfection lies in the realm of fulfilling one’s purpose. In his work Indaba My Children he gives voice to his elder Maynah. we. to destroy his enemies. Such a man [or woman] would not use his tongue to speak. I argue perfection is not an issue of living a life mistake free.traditions empower human beings to penetrate the mysteries of life and to find the appropriate response to its challenges. he or she must be initiated into adulthood. perfection takes on a slightly different moniker. we are nearer this ideal today than most people think (Mutwa 1964: 612). It allows the human being to gain consciousness of his humanity (Zahan 1979: 54). who describes the state of perfection as follows: My son. Aspects of African Civilization (Person. values and skills. A person who doesn’t get initiated will remain an adolescent for the rest of their life. Spirituality. would use the forces of his brain. the chosen ones have been trying to breed a new kind of [human being] who will walk this earth without fear of flesh-and-blood enemies and who could offer defiance to some of the gods who hate us. his mental power. as a progressive passage from exteriority to interiority. So perfection can only be judged by examining the totality of one’s life. Initiation is about maturity. a person comes to earth with a mission to complete. Religion). 5 21 . religion Paris: Présence africaine. The word perfect simply means “to complete a task. To achieve the level of perfection as expressed by Mutwa and others. Zahan says: Initiation in Africa must be viewed as a slow transformation of the individual. Translated by Susan B. The development of the person will take place at the rhythm established by the great periods of bodily development. 5 Centers of wisdom are sacred places where blessings of all kinds are awaited or received from life itself. For instance. each of which corresponds to a degree of initiation. such as good fortune. in his book The Religion. The knowledge of how to receive the gifts of life. for thousands of years. As Amadou Hampate Ba (1972) notes: The purpose of initiation is to give the psychological person a moral and mental power which conditions and aids the perfect and total realization of the individual.” As discussed previously. For Credo Mutwa and the Zulu wisdom tradition. Those who are not initiated into the customs of their ancestors are perpetual children in the eyes of indigenous people. He/she would be able to swim against the current of “river time”—backwards into the past or forwards into the future—as he pleases. culture. Without this understanding of the use of sacred arts and what they consist of.” In other words. dangerous. 1994: 23). Kongo. This work seeks to contribute a general understanding of African symbols. so we will only highlight some important topics that relate to our discourse. To understand African sacred arts. they will become an elder somewhere in their late forties or early fifties. The first-order meaning of a symbol is the primary literal meaning. To accurately speak on symbolism is a whole volume in of itself. KoOnga. Three principle works explore these art objects and are worth mentioning here. This includes characters in African myths which are the focus of this work. etc. and unnatural situation. someone who is selected by their ability to team up with you in the fulfillment of your life purpose. priest. with the singular being BwAnga. however. its pedagogical strategies. person. The maanga are used in the healing arts. In the ciLuba language this can mean “Medicine.” This is the plural form. The study of African sacred arts is critical for examining the pedagogical methods of African wisdom centers. Sacred Symbols We discuss all of the above at length to bring us to this point in our conversation. Paul Ricoeur. Graduating to this new status. The root is present in the words: Ba-KwAnga. If one obediently walks their life path. The art objects under discussion are known as maanga which are symbolic embodiments of spirits that represent and enforce socio-cultural laws called mila. The first is Clementine Faik-Nzuji and her 1996 work Tracing Memory: A Glossary of Graphic Signs and Symbols in African Art and Culture.frightening. and KaAnga. we must first understand the nature of symbols and its importance to the human collective. depends on one’s good track record (Somé. or the types of human beings these centers of wisdom seek to develop. 1999) speaks about this same craft. In Central Africa. characterizes symbols as having a double intentionality. Kalanda Mabika also explores these healing arts among the Baluba of Congo in his work La Revelation du Tiakani (1992). no categorization can embrace all the semantic possibilities of a symbol (Ricoeur 1976: 57). Much of the misunderstanding in the Biblical narratives concerning other cultures is a direct result of the Hebrews having no earthly clue about African wisdom traditions. one will not have the necessary keys to unlock the wisdom that has been encoded in ancient myths. The root of the word maanga is anga which is cognate with the Ancient Egyptian word anx “life. a common term is used to denote symbols in their respective centers of wisdom. a semiotician. which points beyond itself to a second-order “surplus of meaning” that is potentially inexhaustible and that gives rise to “endless exegesis. Also Barbara Thompson in her PhD dissertation titled Kiuza Mpheho (Return of the Winds): The Arts of Healing Among the Shambaa People of Tanzania (University of Iowa. After initiation. From this we can gather that African centers of wisdom make sure that human beings start their quest of perfection during the adolescent years in order that they may stand a chance to achieve perfection during the sunset years of life. These art objects play a fundamental role in the creative process that relies on the composition and performance of symbols and imagery— 22 . the elders will pick up a partner for the young person. It is my argument that one cannot engage this current debate on the borrowings of the Hebrews from the Africans without critically engaging and understanding African sacred arts. oath. remedy = the protected and cared for. Anyone remotely familiar with African initiatic practices can vouch for the multiple levels of meaning for African signs. enable the knower to penetrate the deepest structure of reality. in essence. The verbal symbols are by far best captured in African proverbs. and words and the deeper messages they convey. This is important to note because Africa’s “Bibles” are embedded in their sacred arts. Proverbs are the language of divination. and inspire us through their mastery of the word. Western colonial scholarship have labeled these art objects “fetishes” and have totally misrepresented their essential meaning and function in African centers of wisdom. with eloquent simplicity. help us to better understand the spirit of African symbols as they note: The symbol for us is not abstract. used to communicate with one’s psyche. 2005: 114). personal adornment and litany—to express itself in both local and global terms (Thompson 1999: 7). The visual symbols are the refined objects that are highly valued for their ability to help the wisdom seeker focus on a particular goal. dance. one wouldn’t look for “textbooks. 23 .” one would study the sacred arts (proverbs. These mnemonic devices are used to remind the sage of how he/she is supposed to act. song. They uplift us with their encouragement. all art objects are connected to one or more proverbs: whether they are sung. danced or visual. cultural parallels between the Kongo and Palestine. certain taboos and what in one’s character is needed to meet the challenges of life. music. and the patterns of the universe. The proverbs and wise sayings are associated with any artistic form. which break open the frontiers of life (Kajangu.). As energy and spirits inhabit everything around us.” Maanga and mdw nTr can best be explained in the words of Adama and Naomi Doumbia in their 2004 work titled The Way of the Elders: West African Spirituality and Traditions. maanga is a creative by-product of African centers of wisdom. we are careful to study their representations and behavior. Maanga encompasses both verbal and visual symbology. On pgs 74-75 they. This is how vast amounts of information are maintained over hundreds of years in oral societies because every aspect of their lives are teachable moments which they do not hesitate to exploit. So. We observe the movement of energy. We honor the prohibitions and customs that govern our relations with the surrounding forces. for example. So when one is comparing. The knowledge of what proverbs are associated with these art objects is what’s missing from modern discourse. In ancient times the Egyptians called this method of teaching mdw nTr “God’s words. each dance move is a message and each routine is a story. comfort us with their insights. for instance. African wisdom traditions integrate all aspects of life by borrowing images from living things to assist students of pre-Western modes of thought and being in refining their intuitive abilities. gestures. these artists of verbal energy heal those among us with grieving hearts and troubled spirits. Both visual and verbal symbols integrate all the arts as multivalent sources of knowledge. So if you see a company dancing. symbols. What’s important for our study here is that proverbs are interwoven into all African sacred arts. We receive our knowledge of life from our sacred oral traditions that interpret these symbols for us. We are thoughtful about our own actions. Through story. What is a story in Israel is a living reality on the continent of Africa as will be seen throughout this book. inyanga or singanga which are all terms that designate masters of wisdom who produce maanga. In Africa in general. and music.through sculpture. etc. These mnemonic devices. Mabika describes the maanga as material designs (típàwú) that are filled with material or immaterial charges (tijimba). the cycles of nature. it is an echo of the spirit world. Maanga is closely related to the word nganga. Proverbs or Odus (oracular utterances) are essential to the Yorùbá system of Ifa. carvings. and open. It is the ancestors who used symbols to record and transfer the deepest truths that would enable their descendants to know life to the highest capacity. As well. In order to penetrate the concealed back-view of symbols. and a concealed back-view (what is actually done) to symbols. In other words. Symbols are daughters of experience. symbols for African people inform us about the invisible world for which the physical world is its shadow. Maanga are used as tools so that the brain can use both hemispheres simultaneously. We are sensitive to the conditions of harmony and peace around us. Observing the signs and symbols teaches us about the divine and mystical nature of the universe and our special role in it. balanced thinker. The front-view is the obvious aspect of the sign. Disharmony is a powerful sign from the realm of Spirit that we heed seriously. we believe in an inherent balance in all things. snakes. This awareness serves to keep us present. in the most efficient manner possible. they are tools for problemsolving. all they could 24 . there are two aspects to symbols. something easy to perceive and something not. They are by far the tools par excellence of wisdom in Africa. while the projecting part of an object often represents masculinity. Symbols in their African context record the unending flow of the most compelling events that define life on this planet. All that is visible corresponds to an invisible source. for example. Our openness fosters dialogue with the spirits. We realize that we each contribute through our conduct.. That which is hollow is often a reflection of femininity. symbols enable wisdom seekers to catalyze creativity. They didn’t have the proper lens for which to evaluate the cultural practices of the Africans. Symbols are the most potent objects that possess the power to awaken the creative process. All of creation holds equally feminine and masculine principles.We recognize that all interactions are meaningful. Everything around us contains light and shadow. and tools of creativity. Africans believe there is a creative purpose to life. However. Kajangu argues that there is a revealed front-view (what appears to be done). The shapes of objects reveal to us the meaning of their symbols. I argue essentially that this is the case with the ancient Hebrews and why in the Bible they took every opportunity to take shots at African symbols (i.” There are not many scholars who belong to African centers of wisdom and who can properly “untie” the symbolic knots of the various cultural motifs of African people. Therefore. As a result. where extensive records of historical events. and because of their effort to establish an identity by being antiAfrican. they made laws against the practice and considered them an abomination (thou shalt have no graven images). Symbols are the language of the psyche. As mentioned beforehand. they enable the knower to penetrate into the deeper structures of reality by enlivening his or her gross and subtle senses and activating the capacity to apprehend the truths of existence. the concealed back-view calls for one to stretch his or her imagination beyond itself. African wisdom centers are also art guilds. The various signs and symbols of the universe keep us aware of how well we are maintaining our equilibrium with the forces of life. I attribute this to living a nomadic life. With that said. alert. that life should be lived like a work of art. etc. The Hebrews only had access to the front-view of the symbols. The objective is to develop a more well-rounded. cattle. or scientific phenomena are not as highly valued as you would find in settled cultures in general. symbols assist in aligning knowledge with the creative purpose of life: to crack the code of creativity. all of this while “borrowing” African ideas and trying to change them to fit their cultural world-view.e.). one must be “symbolically literate. when they saw a symbol of a cow. Because of this artistic component. Because the Hebrews didn’t understand the educational value of symbols. The ecstatic arts work hand in hand with other sacred arts to choreograph rituals that enable initiates to learn a few secrets of life. dimona bilengu-lengu (perception of hidden reality).” I highly recommend a book called Boiled Energy: Community Healing Among the Kalahari Kung written by Richard Katz. The two fundamental ecstatic arts are music and dance. The ecstatic arts use sound and movement to assist human beings in the pursuit of wisdom. These arts include stories. proverbs and incantations. “You can tell the health of a people by how well they dance. indigenous modes of thought and being. from my comparative research. These 25 . Because they couldn’t properly code and decode Egyptian symbols. An Amazulu proverb states. The third major arts are the divinatory arts. The verbal arts use the power of speech to assist human beings in the search for perfection. they couldn’t see the association of the human body and how the cow’s head related to the female anatomy. Mabika (1992: 12) identifies four techniques of enlightenment used among the Luba wisdom centers that derive from their divinatory arts: nnduota (visions). There are essentially three major types of sacred arts that express indigenous modes of thought and being in Africa: the ecstatic arts. The Field of Sacred Arts In this section we will attempt to explain how sacred arts embody and display pre-Western. the verbal arts and the divinatory arts. which were invented by the goddess Marimba.” Credo Mutwa (1964: 92) discusses the ecstatic arts.” I think the Kung would agree. A drum would be a tool. “The song is not a song. Mabika discusses the function of verbal arts among the Luba community of memory. for example. the majority of Africa would agree with Mabika. They act as voices of memory for African wisdom traditions. didyongaola (ability to activate one’s shadow or double). divination seeks to open the great gates of eternity in order to find relevant information about its future members. In other words. They use sound to unleash the charge or powers of symbols. However. dancing isn’t just for entertainment. They animate ceremonies and define how rituals are conducted. in the ecstatic arts. and didyalula. An initiated elder once told me that. This demonstrates the caution we should take when attempting to view all African cultures as the same. In other words. but is part of the healing process. These arts go hand in hand. Simply put. We will touch on other associations throughout this work.see is a cow. but which. Kajangu defines sacred arts as webs of symbols that operate as tools of wisdom to assist human beings in the quest to know life and to find appropriate responses to its challenges (2005: 118). conceive and to broadcast graphic images from all living things. The verbal arts allow us to transform the language instinct into a tool that allows us not to lose the powers to receive. This work explores the sacred dances of the Kung of south central Africa that are used to raise energy (num) of the dancer so that he can use it to restore to health to persons who need healing. when released makes one feel closer in the “arms of Eternity. but a movement. He goes on to state: All these dances were invented for one reason only—expression of tribal religion and the release of that beneficial life-force in every human being. nndengama (ecstasies). and informs us that they were designed to assist human beings in unleashing their souls. He argues that verbal arts are the highest form of communication. This is in contrast to Malidoma Some (1998) who posits that speech is the lowest form of communication. they didn’t know it was a sign for doctors (on one level). the philosophy. whose cultural practices go back thousands of years before the historical era. the living institutions and rituals of the people who never left Africa. but in the languages. dream interpretation is still a form of divination. The aim of discussing this in-depth was so that those readers who are not familiar with the study of African cultures and its philosophies could have a foundation of materials to explore in their continued studies of African people. 26 . The divinatory method used in the Bible primarily is what the Luba would call nnduota “visions.techniques are taught by Luba centers of wisdom to assist in achieving higher states of consciousness. With this understanding. As strong as Christians would like to divorce themselves from “pagan” practices. as a lot of the comparative material on the side of Africa will not be in large holy texts. In summary we have analyzed and introduced major aspects of African sacred arts. This also allows us to see what ideas and practices have changed once these customs and ideas left the modern continent of Africa. We seek here to demonstrate how African communities of memory use sacred arts in the pursuit of wisdom.” Practically all revelations were revealed in dreams in the Bible. it better prepares the reader for the journey we are about to embark on. One will need these tools to properly examine Biblical customs and characters because it is out of the creative and spiritual genius of African people that the Biblical stories can be properly interpreted. totally defying current classifications. The anthropological tool that has not been utilized as often is that of historical comparative linguistics. archaeological evidence is essentially inarticulate. I often get asked. and social anthropology. “on linguistic grounds. We start this exposé by comparing the two names (Jesus and Osiris). I haven’t come across in the literature linguistic connections that would assert Jesus and Osiris are in fact the same name: a dialectical variation of each other (like Hebrew Yochannan and English John). We argue here that Osiris and Jesus are the same entity.” Modupe Oduyoye (1984. While many have made conceptual connections between Jesus and Osiris. much debate has gone on for decades on subjects that can easily be answered by examining and cross comparing languages of the Judaic traditions and those of inner Africa. astronomy. Jesus vs. linguistically.IN THE NAME OF JESUS? Many approaches have been utilized to make connections between the Hebrew faith and that of other spiritual systems including archeology. living history of a people’s soul. Language is the oldest living witness to history. with significant differences in their overall narrative.e. In terms of value. Without linguistic corroboration. It is one thing if we find similarities between two stories with primary characters whose names are different (i. the method primarily used is comparative religion which seeks to establish connections based on similarities in themes between the traditions under examination. it makes it harder for Christian theologians to downplay the other parallels that have been made historically and the others that will follow throughout this work. cryptology. These would definitely be grounds for a deeper exploration and harder for Christians to argue against influence or borrowing. but whose connections are undeniable to say the least. like Dr. 1996) has already established the link between Yorùbá and Semitic languages. As a result. for if it can be established that they are in fact the same name. This is what we assert in this section and what follows is the evidence to support this claim. It is because of works like this and others done by African linguists. as an archeological find can often tell you what. The field of comparative religious studies has been severely handicapped as a result of the undervaluing of this tool of analysis. Dionysus). argue that the 27 . to inner Africa that has not previously been explored systematically. Theophile Obenga. that many scholars. It has preserved the inner. I simply answer. The same has been done by Chiekh Anta Diop (1977) between Wolof and Egyptian. When it comes to establishing a connection between Jesus the Christ with earlier religious cult figures. “On what grounds do you have to compare the Hebrew cultures with those of SubSaharan Africa?” Aren’t they geographically miles apart and aren’t Africans and Hebrews different races? Without getting into a debate on the scientific grounds for the existence of biological races. It is another thing all together if the same events happen in both stories with the main characters having the exact same name (like Dorothy from The Wizard of Oz and Dorothy from The Wiz).. linguistics is the next best thing to an archeological find. but comparative linguistics can often tell you why: something that is very valuable in the field of histiography. Many researchers have taken the Biblical text and its common etymologies at face value at their own peril. We seek to understand the true nature of the name of Jesus in this section to establish a connection. in other words. The Hebrew language is a dialect of the Canaanite language. Keita in Bengtson 2008). Kongo Saharan. It should be understood that languages carry the philosophy and values of the people who speak it.7 The Hebrew language belongs to the Semitic branch of Afro-Asiatic. for if cultures share dialectical variations of the same language. This is one of the four recognized macrolanguage families in Africa: the others being Niger-Congo. it is my argument that is it virtually impossible to understand the Biblical tradition without critically engaging African cultures and languages. In these motifs lie the keys to unlocking the obscure customs that are causing so many headaches for researchers in the field. the Canaanites. However. are Africans. The predialectical parent for this language originated in Ethiopia or further up along the Nile Valley (see Bernal 2005. We are hoping that traditional cultures in Africa can give us the keys to unlock the meaning behind ancient Hebrew customs and myths. 7 See Edgar Gregersen (1972). then it can also be assumed that they will share the same or similar cultural practices as well. Khoisan. Journal of African Languages and Linguistics 4:46-56. the Canaanite people belong to the same group of people as the Kushites and Kamites (Ham). according to the Bible. According to the Biblical text. whose philosophical orientation is not African-centered. Based on my comparative studies of Africa and the Near East. In discussing the philosophical phenomenon called Se’en (to be discussed later in our discourse). based on sound comparative methods. By examining the cultures for which the Canaanite languages arose.Afro-Asiatic6 language family is a myth and that new classifications of African languages need to be conducted. An African Writing System (1997: 120) discusses why it is difficult for outsiders to properly decipher Ethiopian terminology and concepts. then we can use the cognate languages to clarify obscure words and concepts in our target language under examination. This is the exact practice we hope to dispel in regards to the name Jesus and the characteristics attributed to him. If it can be established that two languages are in fact related. Ayele Bakerie in his book Ethiopian. he notes the following: The establishment of this language family is in dispute by African scholars on account that no systematic comparisons have been conducted to actually establish the language family using the comparative method as has been done for Indo-European. The Semitic branch of Afro-Asiatic is the only branch of this language group that exists outside of the continent of Africa. This can also help us with understanding rituals. Languages belonging to the same subgroup are believed to enjoy a common period of development and a special interrelation (Campbell-Dunn 2008). It can therefore be asserted that Hebrew is an African language. When it comes to Biblical scholarship. we can better get an understanding of the pedagogical methods and values that have been passed on from generation to generation. The general consensus among linguists concerning the classification of the Hebrew language is that it belongs to a family of languages called Afro-Asiatic. 6 28 . many researchers have fallen victim to inter-Hebrew etymologizing. and Nilo-Saharan. whose results often end in folk-etymology: the attempt to try to make sense of a word no longer properly understood by giving meaning to that word using words that sound similar within their own language with no external verification in related languages. there is a growing consensus in the field that Niger-Congo and Nilo-Saharan are actually one language family called Niger-Saharan (or Congo-Saharan). completion of a task and the beauty of one’s attitude or behavior. 29 . In other words. poetry and/or reading. They also have to be graduates of the schools of music. often giving erroneous interpretations to various cultural phenomena.” I essentially argue for a similar dilemma in regards to the Egyptian and Hebrew cultures—which the latter was adopted by Aryan tribes in the second millennium BCE (Ben Levi in Karenga and Currthers 1986). Because they have never belonged to an African wisdom center. Modern researchers often do not have the conceptual tools to properly analyze African customs and philosophy. the word has both simple and complex renderings that could only be decoded and understood by those who have the skill and the mastery of the Ethiopic language and its grammar. It is my hope that this small work makes a significant contribution in chipping away at the Eurocentric hegemonic perspective that has dominated the fields of both Biblical and ancient Egyptian studies. daginat (kindness). and do not know or fully understand the ancestral vision and pedagogical methods of the people. Se’en evokes both surface and deep expressions and meanings. they often try to force a perspective on African customs using a Euro-Centric framework that is incompatible with African social paradigms. We now begin our discussion on the name of Jesus and our comparison with Osiris of ancient Egypt.Se’en is described as malkaminat (fairness). The ability to disclose the hidden meanings of a given work of Se’en is like untying a given knot and it is fundamentally different from a work of translation. bagonat (generosity). Fr. "relation of incidents" (true or false). Thus יהוחנןYehochanan contracted to יוחנןYochanan. Koso (Kush) One way to demonstrate migrations of people from one place to another is to cross compare the place-names in the regions under examination. However. The settlement called Koso (k-s) in Northern Nigeria has obtained a folk-etymological history. with some advocating a Arabian origin (independent of the advent of Islam)." : from historein "inquire. they will often name the current settlement after the old settlement for which they came. record. Diop 1991) is the old territory of Kush (k-ṣ). The name for Jesus is given as Yehoshua (Old Testament Joshua) and in latter times as Yeshua. not differentiated from story. This is problematic as it is a tell tale sign that this was a borrowed term into Hebrew for which the Israelites did not understand its meaning. This area of the Sudan. who issued an edict that no one should spread the news that Ṣàngó 8 Bearing the name of a deity 30 . Related to Gk. many people believe that the word history is composed of two English words: his and story. Below we will give two examples of this in history: one from the Yorùbá people of Nigeria. history. historia "a learning or knowing by inquiry. Because the common-folk are not comparative linguists. This is quite a common practice in many languages and English is not immune.became Ye-." from PIE *wid-tor-. to understand the terms they use and if a meaning is not readily available. we know in the United States there is a place called Yorktown. sense of "record of past events" probably first attested late 15c. The folketymological explanation is that the name Koso derives from a decree of the supporters of Ṣàngó (the fourth Alaafin of Oyo). they called the new settlement New York. history has the following “history:” late 14c. we know the earliest settlers of that city came from Yorkshire England.E. tale. "to see" (see vision). historie.." from Gk. For instance.FOLK-ETYMOLOGY It is my contention that the common meaning for the name Jesus is folk-etymology." from histor "wise man.to Yo-. When ethnic groups migrate and settle in a new area." and to eidenai "to know. When the British captured the colony of New Amsterdam. It is clear that the history of Yorktown. and the other from the Hebrews in the Old Testament. According to the Online Etymological Dictionary. often. judge. story. from base *weid. History is therefore the “knowledge of records. and even Arabia (Muhammad 2010. idein "to see.” not his + story. Virginia. historia "narrative. from O." In M.." lit. A comparative analysis demonstrates that this is not the case. decided to give it a meaning based on terms already available in the Hebrew lexical inventory."to know. narrative. there is no name (aside from Yehoshua`) in which Yeho. As an example. account. Sense of "systematic account (without reference to time) of a set of natural phenomena" (1560s) is now obsolete except in natural history. then it is a common practice to make one up until otherwise corrected. Virginia and New York city are connected with the history of Yorkshire England. Along with other historical data. This is what happened in the United States as seen above. The Hebrews. from L. The Late Biblical Hebrew spellings for earlier names often contracted the theophoric8 element Yeho. More and more evidence is making itself known that supports the oral traditions of the Yorùbá which state that many of their ancestors arrived in Yorubaland by way of the Nile Valley (Sudan). Many times a whole myth can arise from folk-etymology. in turn. they do not seek. GJK Campbell-Dunn. in west Africa Yorùbá). they analysed the term into two words readily available in the Yorùbá language ko + so. 1984:30). Because the early Yorùbá did not know what the term Koso meant. The writers here attempted to exploit a legend whose date could not be placed in a prehistoric period. 11:7-8] Therefore he called its name Babel For there Yahweh balal [confused] the speech of the whole earth [Gen. Europa. ma-‘areb. Oba ko so “The king did not hang. It is the name of the Babylonian capital whose city gate was memorably designed with religious motifs. Place-names can tell you a lot about a people’s movement throughout history. they utilize a chain during divinations called an Opele (p-l) to ascertain information on one’s destiny. An interesting comparison here is with the Greek god Apollo (p-l-l). This gate came to be known as Babylon: baab ilu “the gates of God” (see Igbo obube and Twi abubuo “door”. the folk-etymologist is basing their etymology on a single leg of sound-similarity. The Hebrew balal means “mingle. This type of design is quite common in Africa as can be seen by this door created by the Dogon of Mali below. It has been demonstrated in the work Who Were the Minoans: An African Answer (2006) by New Zealand linguist Dr. Among the Yorùbá Priests of Nigeria. given factors not mentioned in this text. This myth also provides an etiological (study of causation/origins) explanation for the riddle of the ziggurats of Babylon. Come. 31 . Babylon The internal evidence of Genesis 11:1-9 suggests that this was not originally a sequel to Genesis 10. Verse 5 of Genesis 11 tells us that Yahweh came down to see not what the Babylonians were doing. balala-balala “scatter. The word Babel is not a Hebrew word. The Kushites who migrated westward reestablished their old kingdom in Koso. confuse. confound” (Chi-Chewa-Bantu. This post flood story is now grouped with the antediluvian story of the sons of the gods: two different time periods. Olu “king. human-beings) were up to. 11:9a] The suggestion here is that there is a relationship between bbl and bll (which may be possible). mix. This story also provides a legendary explanation for the origin and diversification of languages among peoples of the earth. but an independent story whose purpose was to provide a diatribe against Babel “Babylon” who allegedly at the time was the oppressor of Yehuedah “Judah” from 596 BCE to 536 BCE. but what the ben-ey ‘adam (sons of Adam.committed suicide by hanging and issued the statement. The more likely history.” (Oduyoye. disperse”). Yorùbá Ehi. A cognate must stand on two legs: form and meaning. god”). Arabic ma-gereb. The historical Yorùbá were Kushites who migrated westward toward the land of the sunset (Hebrew ‘ereb. is that it is a place-name of the old territory from which they came. This is where we get the Hebrew folk etymology of the name Babylon. Greek Ereb-os. This same practice has happened elsewhere in Nigeria as the people of Oyo fled from Oyo Ile near the Niger to reestablish the kingdom of Oyo north of present-day Ibadan. that one of the earliest settlers in the Aegean was in fact the Yorùbá people. However. the god of destiny. let us go down and boll-ah [confuse] their tongue there…[Gen. If we could discover this meaning we would have a good idea of the original character of this god. as it is a part of the name Jesus. not fact. YHWH The first term under examination is the word YHWH. a whole mythos has been created on the misinterpretation of a foreign word.10 Therefore. 9 Therefore. have failed to supply a convincing solution to the Retrieved from. As we can see here. Robert H. who have suggested numerous etymologies of this name. Pfeiffer relays the following: The name Yahweh in the Semitic dialect spoken by the Kenites must have had a meaning and thus characterized the deity.Spiritual motifs on a door from Mali of the Dogon people West Africa. This.5cense. But the efforts of modern scholars. To the writers of Genesis. I contend. There is a lot of folk-etymolization in the Bible when it comes to foreign loan words. The folk-etymology given by the Hebrews is that it derives from mashak “he drew out” (Exodus 2:10b) to fit the myth that Moses was taken from a river and adopted by the house of Pharaoh. However. it only serves as a means to end. is ultimately the case for the meaning given to the word Jesus/Yeshua. Babel could not have been so called because God “confused” the tongues of man. we cannot take the Hebrews on their word about terms without resort to comparative linguistic data from African languages. 9 10 32 . Although this myth produces a great story. Accessed 2/3/2011.” mèsú “son”.htm. The name of Moses (Hebrew Mosheh) is the most common. a word in a strange language must connote something similar in a similar-sounding word in Hebrew. This end was simply to try to explain certain phenomena and situations affecting the world around the early Hebrew people. Egyptian msw “begotten (of)”). The etymology given in the Bible is based on fancy. the name given as Moses is actually an Egyptian word msi meaning “he gave birth” from the root ms “child” (Yorùb| mèsì “born. 52-53 (cited in Saakana 1991: 62-63) " לכֵן יִ תן אֲ דנָי הּוא לָכֶ ם אוֹת הנֵה הָ עַ לְ מָ ה הָ רה וְ ילֶדת בֵ ן וְ קראת שמוֹ עִ מָ נּואֵ ל ְ ָ ָ ֶ ֹ ָ ִ ֹ ֵ ָ 13 Among the Ewe. This is an interesting passage given that in Matthew 1:23 it states. West African languages may offer a solution to the original meaning of the terms for God in the Bible. a goddess who sometimes impersonates men. and you shall call his name Jesus. they have had to do a lot of guesswork trying to understand concepts that are common place in Africa. The word belongs to a lost language. As a result.” Matthew 1:21. Another etymology posits that Yehoshua derives from Yahu “god” and shua “a cry for help. 1984: 98): Hebrew ‘el ‘eloah ‘el šadday ‘el ‘eley-own ‘ele-yown YHWH Divine being God God Almighty God Most High Topmost Yorùbá Elu/Olu Oluwa (-l-w) Jukun Tsido Ibibio Abasi Enyon enyon Ewe. and Avle. A few notable ones are Heviesso. and they will call him Immanuel"--which means. master God The Supreme God “peak” “the power in lightning and thunder” The Hebrew term given as YHWH (Yahweh) is given in Fon as Yehwe “spirit”. a saving cry” (Tshiluba-Bantu swii “cry”). therefore they have not sought out solutions in understanding Biblical names and cultural nuances by examining these foundations on the continent of Africa. In other words.11 Not much headway has been made since this publication in mainstream Biblical studies. pp.” The Yeho. Nor does Ex. scholars aren’t sure what the names YHWH or Jesus means. for he shall save his people from their sins. The idea is that with the two terms together it means “God is a saving cry” (shout to God when in need of help). It appears that they are trying to force a definition on the name Jesus based on the following passage: “And she shall bear a son. Fon Yehwe God Lord.mystery. Ewe Yeve “spirit”13 and in Gun Yihwe. There has been an attempt to link the name YHWH to a Semitic stem HWH (originally HWY) meaning “being” or “becoming. In essence it refers to a shout given when in need of rescue.” There may be some merit to this as we will see later on in this discussion. This is primarily because they do not think that the Hebrew tradition is linked with inner Africa in any meaningful way.” These are not loan words from some benevolent wandering Religion in the Old Testament (1961). "God with us" (citing Isaiah 7:1412).name is alleged to be a contraction of the tetragrammaton YHWH ()יהוה which has become the proper name for the Old Testament god. the Yehve spirit is really a class of spirits. For now. "The virgin will be with child and will give birth to a son. 11 12 33 . Harper & Row. 3:14 supply a clue to the meaning…None of the sources of the Pentateuch had the slightest inkling of what ‘Yahweh’ meant originally. Yehwe “God. god of thunder and lightning. it is believed that Yehoshua/Yeshua is composed of two names Yeho “God” and Yasha “save. including YHWH (Oduyoye. As discussed earlier. The two different etymologies do not instill confidence in the proposed meanings. The later Christian writers attempted to use this verse to find legitimacy for the virgin birth story of Jesus. “Can every instance of the term YHWH in the Bible be translated as God?” In Genesis 4:1 it states “I have acquired a man ‘et YHWH. The qowl of YHWH is in majestic splendor. As shown above. “You will then invoke your god by name. 1 Samuel 7:10 The feminine form Balat in Ethiopic is a term for gods or goddesses of fire or the sun (Bekerie. that qaniy-tiy ‘iyi ‘et YHWH. To answer these questions we will examine the Yorùbá deity Ṣàngó. “I have acquired a man and God”? The interpreters of the Bible who translate this passage as “I have acquired a man with the help of God. Did the Hebrews recognize YHWH in this same manner? It appears so.” Could Hawwah (Eve) be saying. The Hebrew word qowl (q-l) means “voice” (Arabic qawel “voice”). 2:13. A rumbling goes out of his mouth.” Could this be the result of a lightning flash? As the record shows. she was comparing the fire in the forge to the fire of lightning. 34 . The resultant version would pose a problem in meaning. and his ‘owr reaches the ends of the earth. 1984: 98). Let us agree the god who responds with fire is the God. There may be an indirect correlation between YHWH and lightning in Kings 18:24. 1997: 70). This would not be the case if they did not assume that the word YHWH always meant “God” instead of “divine spirit. These are inherited terms. 17:13).Hebrews who decided to teach West Africans the name of their God. YHWH of the Bible is an ancient deity that represents the power and the spirit behind lightning and thunder. who like Yahweh of the Hebrews. and I will invoke YHWH by name. The question becomes.” Yehwe among the Fon and Ewe represents the power of “thunder” (and lightning). After it he booms his qowl and thunders his majestic qowl. all of the Old Testament names of God are reflected in this area of West Africa. 14 15 See also Psalm 18:14. Another picture of Kalu (qowl) in the Hebrew text can be found in Job 36:30. The qowl of YHWH breaks the cedars – YHWH causes the cedars of Lebanon to be broken. wouldn’t it be common to associate this deity also with rain/water? Is there any scriptural evidence to support such a supposition? The Biblical text states that God himself can be called "the spring of living water" (Jer. The Biblical character Qayin (Yorùbá Ogún) is the patron of blacksmiths and farmers. 37:5:14 Listen to the boom of his qowl. He lets it roll under the whole sky.” are not translating the Hebrew word ‘et “and. 2 Samuel 22:14. YHWH God of Rain? If YHWH is a god of thunder and lightning. We will see later how all of this begins to make sense in the context of the figure Jesus of the New Testament. In Psalm 29 it states: The qowl of YHWH is (heard) upon the flooded waters – ‘el of Kabowd causes thunder… The qowl of YHWH is in power. with” (Yorùbá ati “with”) or the particle preceding the direct object (Oduyoye. When Hawwah mentioned that her blacksmith of a child was a YHWH (Yehwe). is also a deity of thunder and lightning. This spirit among the Igbo of south eastern Nigeria is known as Kalu (k-l) whom you hear in the crack and boom of thunder. The qowl of YHWH makes the lightning flash. in regards to the birth of Qayin (Cain). Here Eliyahu tells the prophets of Baal15. Egyptian Twi Yorùbá Lugbara Mbuti ciLuba Gurma Gurmantche Fulani Masai Kwasio Mombutu Ewe Ijo ntr ntr ntoro ntori adro Adro Ndura Ndele Unteru Untenu Ntori Naiteru Nture Noro Tre Toru “natron” (cleansing agent) “God” (unseen fructifying agent) (Coptic noute) “spirit of patrilineage” “because” “guardian spirit” “God” (also the whirlwind found in rivers) 16 “God” (<of the rainforest) “divine. It is used as currency and it is a symbol for wealth and blessings. begetter. As the Igbo Kalu is qowl YHWH. Facts on File Note that Budge (1008a) has a rendering of Ntr “temple of Isis” with the water canal glyph as a determinative. Aje Saluga is the god of wealth. to be pure” “rectitude” “honesty” “shells” “sun” “beads” “divinity of lighting.The word Ṣàngó is built on an s-n-g root which has a by-form in Yorùbá s-l-g: Aje Saluga which is the cowrie shell that has been deified for its pure whiteness. Hebrew ma-tar tal is “rain” (cleansing agent. Among the Azande of central Africa we have. African Mythology from A-Z. so is the word of YHWH compared in Isaiah 55:10-11 šeleg and gešem “rain. thunder and rain” Here we see among the Dinka of Sudan their god of lightning and thunder is also associated with rain.” Both of these elements are fructifying and purifying agents. In the Hebrew language there is a reflex in the terms seleg. This root occurs in the following African languages: Duala Kikongo Eko Acoli Kiswahili Dinka sanga n-songa n-songi o-sing ceng u-sanga Deng “to be clear. and other African nations.” Biblical Aramaic telag (s>t) “snow. Ancestor” “God” “God” “God” “God” “sacred” “God” “clan spirit. Hebrew thr “be pure”) “dew” (uncontaminated water) The ancient Egyptians. fructifying agent) is “God” “rain” (cf. toro Ma-toro cf. saleg “snow.” both of which is pure white. 16 17 35 . fetish” “river” (Egyptian i-trw “river”)17 See Patricia Ann Lynch (2004). This association of purity as symbolized by rain (water) is at the heart of one of the ancient Egyptian terms for the divine: nTr. associated their gods with the purity and the life causing essence of water. . spit PWS la. Erasmus University Rotterdam.3. Alarodier und Proto-Phrygen. Tyrrhener. ‘God. & S.) : *TVKV ‘to pour.J.) : *tokʷ ‘saliva. AnTari . NTr qbe . acted flamboyantly. 18 36 . refined gold. 245f. (da) “day” (sky = rain) PWS gi “firmament” PWS na “above” Mangbetu ro “sky” Sumerian an “sky”. Leiden / Philosophical Faculty. werq anTari . The more plausible route is from Africa:19 Connections proposed by Karst20 Mediterranean Canaanitic El / Bel. In other words.” I do not find this etymology tenable. the sky is symbolic of the infinite vastness of the Creator. Liguro-Leleger.Tonga Amarigna18 Tilo AnäTära “blue sky. ProtoAustroAsiatic *tVk ‘drop’. rain. Remarks Proto-Bantu *d often changes into – l. mlungu. ‘God’ Meeussen. It is also a metaphor for the source of life generating essence as can be observed by rain falling from the heavens which impregnates the earth and brings forth crops (Oduyoye 1996.refined butter.und Ostiberer.n. Tnte nTr . Heidelberg: Winters. mulungu. [ >-ilu. Ford 1999).da “rain” Gur *do “sky” North Guinea (Johnston) dana “sky” [a + a > ā] Mande la. goldsmith.t. refiner. Origines Mediterraneae: Die vorgeschichtlichen Mittelmeervölker nach Ursprung. It has been posited that the word god derives from an old Indian term huta which means “the one who is invoked. Etrusker und Pelasger. to rain. refined. Schichtung und Verwandtschaft: Ethnologisch-linguistische Forschungen über Euskaldenak (Urbasken). Lyder und Hetiter. with Guthrie number -*gòdò 5-. AnaTere . had refined. cf. Heaven’ muluku/m-luko. drip. drop’ (> Eurasiatic: *tUKV . van Binsbergen’s (2009) paper "The continued relevance of Martin Bernal’s Black Athena thesis: Yes and No. Jolos. God” (from which the rains fall) “pure” In the vast majority of African spiritual systems. The name for God is not a Germanic word as some linguists like to assert. sky. AneTere v. bounced. God is a title for our oldest and greatest ANCESTOR. AnaTara .C. Jolaos (j-l) Phoenician/Punic Moloḫ Bantu y-ulu.." African Studies Centre. above. Heaven’ Proto-Bantu Guthrie.. AsneTere v. Amerind (misc. CS Lendu ra. Lotuko (na)are “water” CN Berta ro: “rain”.n. [ > S. 880. 20 Karst.n.n. AnT'renya – silversmith. 6. top. was refined. 650. West. 1931. sky.t. Austric: ProtoAustronesian *itik. Mangbetu kini “night” Bantu bú. J. arra “sky” This word ntr is reflected in Amarigna in the following terms: NiTir .v.4. ManTeriya .in historic attestations. Pyrenaeo-Kaukasier und Atlanto-Ligurer.i. Ntr werq . rain”. *Borean (approx.pure 19 See Wim M. Tama ar “sky.refining flux. Dinka uar “river”. 5.C.adj refined.i. teneTere v. Sardinian /Aegean Julus. & -dÓk-. -dók-. with noun classes -*gudu 5 L LH. wilu ‘God. (da) “day” ES Afitti araŋga “rain”. Sino-Caucasian: *[ṭ]Hänḳó. e-ulu.in S. Element (chemical). > ? Mu-lungu. pp. Bari ko. The sky is only used as a metaphor to denote the highest example of excellence.prefix. Mw-ene. Igbo enu “top of. and -d-21.” Olánrewájú “status is progressing forward. height. the distance (in ability. òrìṣà. etc. fame. /r/ or an /n/. buy.” Oláitán “honour never gets used up. the head honcho. Bw-ena. God.ŋ. -n-. lord. and Unkulukulu (the oldest of the old) among the Amazulu.” The root of Hrw is wr/ulu (Yorùbá Olu “great.” san-kolo is “heaven”. etc. Bambara san “rain. the eldest (the oldest thing in existence). honourable estate. NigerCongo k-ala > Arabic ‘l). Bagirmi kada “sun. distance.).” It is from this variation I believe that the word Allah (‘l) ultimately has its origin (which originally had a k. the peak of something.prefix that became a glottalized stop. the most honourable. sky. God”. anything that is tall or reaches the heavens can be an #ulu (sun. up”). P-ala/H-ala. exaltation. Gulu among the Chagga-Bantu of East Africa (d>l). Mangbutu kora “day” CN Didinga kor “sun”. enu “top. niombo.lo. These terms evoke a sense of “eldership. day”. Babylonian Ilu “God”. REMOTE Sumerian ul “be distant. stars. K-ule. 21 37 . K-ulu.or g. Z-eru. In Ancient Egyptian this term became Hrw “sky. an apex. 22 The vast majority of African traditions do not view the Creator as intervening directly into the lives of humans.” For this reason. Dinka ako. for instance. angels. Z-ulu.” Eloh’iym “Gods”. El “God. the summit of achievement (what a priest represents). Often the k. sun. 22 Campbell-Dunn (2009b) demonstrates this with the following terms which support our thesis above: DISTANT.or gprefix will go through a process of palatalization and it will become an /s/ (Akan O-s-oro “sky. A related Yorùbá word Olá means “elevated status. Turkana ekoloŋ “sun” (Greenberg) PWN NI “rain” (from on high) Bantu niin “ascend” Swahili nya “rain” Mande yire “ascend” Mande mi “rain” The Nilo-Saharan. Kulu among the Bakongo.l “sun”. ntrw. remote” KU “big” PWS ku. thus why Africans call on his agents: abosom. Niger-Congo consonantal root for sky/heavens/God is -l-. high”.” Ol|dŭnní “high status is sweet to have. It is this root from which we get Yorùbá Olu. Arabic il’ilah (Allah) “God.CS Kreish kadda. Hebrew Eloah “God.” In Bantu this term is rendered G-udu. etc. Specific examples of the name God in Africa can be seen in the name Guéno (the Eternal) among the Fulani (l>n). (n)K-ale. god”) or a /z/ (Amazulu Z-ulu “heavens. Canaanite El “God”.” It can be seen in such names as Olásení “fame is not unachievable. often with a k. the /l/ root (ilu) in another language will be a /d/. as in the famous Tshaka Zulu). Elu. year. G-ulu. So in one language. consciousness and wisdom) between man and the Creator.” Oluwa “lord”. moon. sky”. the possessor of all things (as the sky encompasses all things in the universe) and general absence from. mountains. kul “old” PWS kua “road” PWN KUA “go” (from *kula ) PWN KWUL “be big” Bantu kúdú “adult” Bantu (Meinhof) kulu “big” -l These sounds are known throughout world languages to mutate and interchange with each other. Swahili maji “water”. we note the following: WATER GI. gia “water”. (bā-n). Sukur (3) yiam. bãg. which in Niger- 38 . In terms of Ea (Yah). The moon. e4 “water” The Sumerian is closer to the Mande rendering and suffers from erosion of initial consonants leaving only the vowels. The name Ea is of Sumerian origin which consists of two signs signifying “house” and “water. Ngala (2) am. however. Poto mai “water”.a “muddy water”. There may be a possible loss of a locative class prefix. This god is also pronounced as Ea (Yah) (Darkwah. bàt “swamp”. Kpelle ya “water”. is connected with the rising tides which is still associated with “water. Kondugr oŋgul “road” CS Bulala kori. Ward (1935). 2009b) has presented a convincing case for a Niger-Congo origin of the Sumerian language based on a comprehensive linguistic treatment of the languages.” Ea was a water deity primarily associated with water underground. Efik m-bat “swamp” PWS pat “swamp” Bangi. It should be noted that Yahweh in Psalms 68:4 is given as Iah/Ieh. Mampa yi “water”. gyi su-ma “fresh water” Mangbetu kuma “rain” (suffix ?) Afro-Asiatic : Chad : Angas (1) am. Kongo maza “water” Mande bã. PWS bà. The Egyptians had a deity named Iah as well.Bantu buk “go away” Mande ku “return” Mangbetu eku “to return” ES Dilling okul. which later became Yahweh among the Hebrews (see also Gnuse 1997: 74-87). Kele balia “water”. flow” [In Sumerian y > #] *M = # *A = a *A = e or *B = # *R = # Sumerian a. These are dialectical variations of the same ancient word describing the same divine presence which has survived in Indo-European as “God. “go” (see previous) Niger-Congo ma “liquid” prefix. 2005: 217). PWS gi. William Durant (1935: 310) argues that the Jews adopted one of the gods of Canaan.” It has been convincingly demonstrated in the works of Darkwah (2005). Mende yia “water” > ye. and Ankaba (2010) that the ancestors of the modern Akan of Ghana once lived in the land of Canaan. gõri “road” Khoisan /Nusan (S) !nu “foot” [Sumerian has lost initial k and final a] *K = # *U = u *L = l or *B = # From this evaluation it can be stated the so-called Semites and the Africans worship the exact same god (literally) and the Hebrew notion that others with these variations of the same name is somehow worshipping another “god” is unfounded and rooted in ignorance of African languages. Ngala. but was associated with the moon (also with Djehuty and Osiris).” Campbell-Dunn (2009a. Yah(u). presumably from *ya “go. GIA “water” RA “water”. Musgu yem “water” (Greenberg). canal. In Yorùbá the notion of “life” and “existence” is given by terms that have a j/y. he lived through (cf. One suspects such homophonous vowels have lost initial consonants. In African cultural systems. 23 39 .interchange (Oduyoye. he is alive” (cf. which is the most used term in Hebrew for “to be. which in Hebrew would be jah/yah. and pa. flood. See CampbellDunn (2009a: 12). house. existence.meaning “place.” All flowing streams (living waters. Aramaic hawah “to be”). “goddess of all rivers”. 1996: 104): O jé jí èjè Òjè “He was/He is – he happened to be” “Wake up from sleep (back into consciousness and life) “blood – life blood” “Dead come back to life” The cognate of jé in Hebrew is hayah. Ye-moji near Ijebu Ode.Congo is typically ka-. *gia “water”. which is achieved in Yorùbá by a j/y alternation: Ó yè ìyè ayè ìye ìyá aya “He survived. This is why in Yorubaland all of the rivers are named after goddesses (Oshun “river Oshun”.” This is reflected in Yorùbá as Ó wà “it exists. nominalising suffix”. Among the Yorùbá the great mother goddess is known as Oya. waters of life) are seen as female among the Yorùbá (and Africa in general). This root is also in the name of another Yorùbá goddess Yemoja (ye “mother” [of] m-oja“water”.” This may explain the Sumerian sign denoting a “house” and “water. tears. this term has many homophones that should be noted: Thus a. Middle Egyptian iw. means “water. def. salvation. Oya “river Niger”.” This term is related to hayyah/hay meaning “life” in Hebrew and Arabic. there is. Ogun [who is also a goddess] “Ogun river”). ku-. alas. seminal fluid. father.” This root is in the mythological creator of the Yorùbá Odudúwà ‘Oracular utterance However. offspring. (e) in Sumerian. earthly existence. article. in. Igbo iwu. water is typically feminine and often is associated with motherhood.”23 More investigations into this question should provide greater insight. Òjè “living after death”) “mother” (the one who brings life) “mother” “wife” (another role of a woman) The goddess of Thursday is Yah (the masculine god for Thursday is Yaw) among the Akan (Darkwah 2005). Remember the m. èjè “life blood”) “Life. life” (cf. world” (cf.” so the root is ja “waters” from PWS *gi. life”/”manner of being. Kiswahili maji “a spring of the water of life”). There is a subtle semantic shading in which Hebrew achieves through a slight alteration in the guttural articulation of the first consonant (ḫ/h=zero in Yorùbá). The verb wà is the verbal base for the noun ìwà/ùwà “being.is a noun-class prefix denoting “liquids. locative suffix “where. character. Could Yahwe/Yahve have originally been a “goddess?” We mentioned earlier that there has been discussions linking the name YHWH to a Semitic stem HWH (originally HWY) meaning “being” or “becoming. manner of existence. This term can also be interpreted as the “source of life. O jé “He was) “survival. when”. ha-. is short for túndé which means “re. This is why there are no female apostles.27 No matter how we look at it. begin. ùwà “life”.” “wind” and is symbolized by a “tornado” (Neimark 1993: 125. having both male and female aspects. In most world traditions. However. Also. ṣè “initiate. Female god and divine Child (usually a male). he is alleged to return and do the same thing. existence” having two roots (which is common). then Yahweh simply would mean “the mother/source of existence.” This term survives in Hausa as uwa “mother. It is also in the name Túnwàṣè “Begin life again. The final Egyptian r sound was dropped often starting in the Middle Kingdom. heaven. further reinforces the dominant spirit of maleness in the Biblical text.” But it also can mean “until” like in the greeting Ó d’ {|rò “until morning.[odu] created [d]24 life [ùwà]” (life was created by the word of God). but the depiction that is most dominant in Western literature is that of both of them as male. cloud. when they talk about “fire stones. disciples or angels in the Biblical text. this time using “fire and brimstone” from heaven (lightning?). This is apparent given the fact that he chose to annihilate humanity by causing a great storm which flooded the whole earth and drowned its inhabitants in Genesis 6: something a rain god would do. among the Yorùbá. 24 40 . This verb d/di generally means “to become.” Eyheh ashar Eyheh The Hebrews are a patriarchal people and they suppressed many divine or positive aspects of women in the Bible. eternal.” Ó d’ òla “until tomorrow.” Wàṣè Yorùbá is Washil in Tshiluba and Wsr in Egyptian: Wsr/Wsr being the regenerative force in nature (the life cycle). “Because she was the mother of all life. in the Hebrew tradition. living” which is a name given to the first mother according to the Hebrew myth. thunder and lightning” is in keeping with ancient African tradition.” If Yahweh is a conjunction of Ya(h) “mother. rain. So there is no conflict in Africa as to whether Yah is a “male” or “female” deity. sky”). Genesis says. Among the Yorùbá of Nigeria and the Bakongo of central Africa. This same tradition gave us our modern term for the divine creative essence commonly called God (*godo “heaven.” This may have some support (minus the feminine association of course) in the Biblical text when God revealed his name to Mosheh (Moses) and states Eyheh ashar Eyheh “I am that I am” which can be further interpreted to mean “self existent.” The fact that Yahweh of the Bible is associated with storms.”25 The Hebrew name Hawwah has come to us in the English speaking world as Eve “life. lightning and thunder.” Ó d’ {bò “until (my/your) return. YHWH is associated with rain (water) and lightning.26 and is depicted as an aggressive god who kills those who disagree with him. life” and Iwa (Hawwah) “life.” 25 This name Túnwàṣè is also cognate with Wsr as the word tún. return” (the dé means “come”). the goddess Oya is also the goddess of “storms. 26 Remember that God in the Bible revealed himself as a “cloud” by day and a “fire” by night (in the heavens) during the Exodus (Exodus 13:21). again. 27 A similar case can be seen among the Yorùbá where the gods Ògún and Eṣu are depicted as male and female. the “trinity” consists of a Male god. the feminine principle has been reduced to a diminutive form of an ambiguous/anonymous “Holy spirit. to cause.” they are talking about “lightning. In relation to the root Ya in Yahweh/Yeve/Yehwe. In many continental African cultures. It just depends on what region you go to and what function of existence the people need it to be at that particular instance. God is seen as androgynous.” “raging rivers.” God being associated with the “sky. re-initiate existence. ” As Epega notes concerning Oya: In her lifetime. she was brave and warlike as her husband [Ṣàngó]. Yahu or Ya. Another point of debate is the nature of the /h/ value for the contraction of Yahweh to Yah. or "to be victorious" In the Torah. This may lend interpretative value for Revelations in that after the chaos of YHWH’s return (the storm) he will give forth his living water (Rev. (Epega. "salvation". The last source for much discussion is in the meaning of shua in Hebrew as it can have two meanings. 21:5-7). 2003: 77) In African myths. It is my ultimate belief that the name Yeshua is a borrowing and it is again in Africa that we can get the complete range of meanings associated with the root. Most of the discussions I have read primarily concern the vowel sounds for which. She would accompany him to the battlefield and fight as fiercely as a man. It should also be noted that the word ìwà in Yorùbá also means “character” and “beauty” as can be seen in the word ìwàpélé “gentle character.ישעYasha .Epega 2003: 76-78). it is a poetic way to denote a union of concepts (a marriage) that when they come together they produce Z. ה ֵׁש עַעHoshea ֹו Greek does not permit masculine proper nouns to end in a vowel. "to deliver/be liberated". Below are the names of Jesus (with the Hebrew characters read from right to left): Jesus in Greek and Hebrew Ιησοσς or Iesous: *jesu-os → [jeˈsuːs] י ֵׁשּועYeshua‘ ַ י ֵֹוׁשׁשּועYeshua` ַ ֻׁ י ְהֹוׁשעYehoshua (Joshua).” Oya and Ṣàngó are the essential elements that bring forth rain (needed to sustain life on earth). So Oya and Ṣàngó are two sides of the same coin: in this essence. She is also the favorite wife of Ṣàngó.” One could argue that Yah (Oya) + Hawwah (Ìw{) simply means “the essential character of storms/water. thunderstorms (the cause of rain). We see in the Old Testament why Yahweh is such a “wrathful” and “vengeful” god: he has the characteristics of storms. She was also fond of using charms and magic. the name Joshua’ (Jesus) is spelled in different ways which can be seen below: 41 . It is this general term that later became the proper name of the divine among Hebrew people. Yasha or Shua? There has been much discussion on what is the “true” pronunciation of the name Jesus in its Hebrew context: Yeshua’ or Yehoshua’. when it is said that X married Y. ultimately. the god of “thunder and lightning. modern Hebrew people do not know. One cannot ultimately look to Hebrew for the meaning of the name YHWH as it was an adopted descriptor of the Creator from a people who migrated out of Africa into the Levant. We will be primarily concerned with the latter question in this section as it relates to the meaning of the name Jesus. ַ ֻׁ . shua/shuwa. It can also be a “v” sound in modern Hebrew.” As we can see from the Old Testament texts below. it is more than likely that the Aramaic form Yeshua was how his name was pronounced. 42 . The fact that this spelling exists in the scriptures proves that the Messiah's name cannot be "Yahusha. 28Table from:. This is why in the Greek the root is Isou with the long ‘oo’ sound after the s. This is the ignored letter in the pronunciation "Yahusha". definitely giving the long ‘oo’ sound thus giving us shua." Silent without a vowel point but indicates an "ah" sound at the end of "Yahushua". produces an "oo" (u) sound as in #7307 Ruach. 3:21 and Judges 2:7 actually gives us another "waw" after this letter. With that said.html.The pronunciation of the letters in the name Jesus can be seen in the table below: Hebrew characters in Jesus name28 יYod הHeh וWaw שShin ו Produces a "Y" or "I" sound. The # sign followed by the numbers are references to lexical items in the Strong’s Hebrew Dictionary. The argument is whether the second word in the name. In Arabic the name is Issa. Waw עAyin The pronunciation of Jesus’ name would not be so much of an issue if we actually had documentation in Hebrew of the New Testament. the waw sign sits in between the shin and the ‘ayin sign.com/t/yahushua. At this period. the primary lingua franca of the region was Aramaic and it can be demonstrated that this is the language Jesus more than likely spoke (if he existed at all) as Aramaic words survive in the Greek New Testament. This is not the case as our only record is in Greek. Also called "Vav" in modern Hebrew.blessyahuweh. proving the "shu" pronunciation as valid. This also eliminates "Yasha/Yahusha" and "Yahoshea/Yahushea" as being possibilities Again. This is why Strong's 3091 gives 2 possible spellings (see above lexicon graphic). As a Hebrew vowel letter it can produce an "oo" (u) sound like in #7307 Ruach. The following "oo" sound is indicated by a vowel pointing but Deut. derives from the term yasha “to save. See Bibliography at the end. Produces the "sh" sound. As a Hebrew vowel letter it can produce the "Ah" (like in #8283 "Sarah"). wealth. the son of King David:--Elishua. and all the days of the elders who outlived Joshua י שועwho had seen all the great works of Yahweh which הו He had done for Israel. see SH8173 43 . see SH1323 see SH7771 see SH1339 SH4444 4444 Malkiyshuwa` mal-kee-shoo'-ah from 4428 and 7769. Shua root in Strong’s Hebrew Dictionary of the Hebrew Bible SH474 474 'Eliyshuwa` el-ee-shoo'-ah from 410 and 7769. saying. an Israelite:--Malchishua. see SH4428 see SH7769 SH7770 7770 Shuwa` shoo'-ah the same as 7769. The numbers are for categorization purposes. God of supplication (or of riches). see SH410 see SH7769 SH1340 1340 Bath-Shuwa` bath-shoo'-ah from 1323 and 7771. Bath-shua. 'Your eyes הו have seen all that Yahweh your Elohim has done to these two kings. Shua. a Canaanite:--Shua.Deuteronomy 3:21 "And I commanded Joshua י שועat that time. Elishua. so will Yahweh do to all the kingdoms through which you pass Judges 2:7 So the people served Yahweh all the days of Joshua. Malkishua. Shuah. Shua. pleasure. enjoyment:--delight. see SH7768 SH8191 8191 sha`shua` shah-shoo'-ah from 8173. an Israelitess:--Shua. daughter of wealth. The following table contains the definitions for shua’ and words used to mean “save” in the Bible using Strong’s dictionary. king of wealth. see SH7769 SH7774 7774 Shuwa`a' shoo-aw' from 7768. the same as 1339:--Bath-shua. Jesaiah. sure. (let. (that lay) wait (for). be circumspect. nourish up. help. regard. the name of ten Israelites. X certainly. repair. properly. from 3467 and 3050. rescue (literal or figurative. suffer to) live. Jeshua. the Jewish leader:--Jehoshua. to be open. to revive:--keep (leave. Jehoshuah. save life. preserve. take heed (to self). causatively to revive:--live. health.e. keep(-er. see SH3467 44 . Joshua). SH3470 3470 Ysha`yah yesh-ah-yaw' or Yshayahuw {yesh-ah-yaw'-hoo}. the name of seven Israelites:-Isaiah. Jeshaiah. preserve (alive).):-deliverance. aid. be whole.SH1340 1340 Bath-Shuwa` bath-shoo'-ah from 1323 and 7771. to live. preserve. 2421). to protect. causatively. defend. victory. save. Jah has saved. to hedge about (as with thorns). the same as 1339:--Bath-shua. observe.e. national or spir.(by implication) to be safe. victory. Jeshajah. Joshua. revive. i. lives). look narrowly. self). generally. also of a place in Palestine:--Jeshua. i. properly. see SH7768 see SH3467 SH3467 3467 yasha` yaw-shah' a primitive root. to live. mark. give (promise) life. prosperity:--safety. avenging. help. pers. deliver(-er). Jehoshua (i. (abstractly) deliverance. save(-iour). saving. see SH2331 see SH2421 SH2425 2425 chayay khaw-yah'-ee a primitive root (compare 2421). i. prosperity:--deliverance. life. see SH1323 see SH7771 see SH1339 SH3091 3091 Yhowshuwa` yeh-ho-shoo'-ah or Yhowshua {yeh-ho-shoo'-ah}. salvation. see SH3091 8668 tshuw`ah tesh-oo-aw' or tshuah {tesh-oo-aw'}. whether literally or figuratively. make) alive. watch(man). save (self). salvation. salvation. daughter of wealth. (X God) save (alive. guard. Compare 1954.. to free or succor:--X at all. from 3467. something saved. causatively. rescue.e. recover. Bath-shua. saving (health). X surely. reserve. quicken. welfare. hence. safety. see SH3068 see SH3467 see SH1954 see SH3442 SH3444 3444 yshuw`ah yesh-oo'-aw feminine passive participle of 3467. liberty. he will save. help(-ing). bring (having) salvation. be safe. see SH2421 SH3468 3468 yesha` yeh'-shah or yeshai {yay'-shah}.e.:-beward. from 3068 and 3467. SH2421 2421 chayah khaw-yaw' a primitive root (compare 2331. deliverance. wide or free. get victory. from 7768 in the sense of 3467. attend to. restore (to life). see SH3467 see SH3050 SH8104 8104 shamar shaw-mar' a primitive root. 3442. Jehovah-saved. see SH3467 SH3442 3442 Yeshuwa` yay-shoo'-ah for 3091. etc. see SH5337 SH5337 5337 natsal naw-tsal' a primitive root. The failure in getting to the root of the name is in not understanding the “root metaphor” for which the name (and all of the other associations derived). throughout the myths in which he is a part of. X without fail. It is my contention that it is the latter. God does not judge. The Nature of Characters in Myths Mythological characters are generally an amalgamation of characteristics. so the name could have been pronounced differently for different contexts. Because Ogún belongs to the g-n/k-n root. In typical African (especially Egyptian) fashion. rescue. It is my contention. but Jesus. part. X surely. that he is not God and never claimed to be. but has entrusted all judgment to the Son. strip.” My objection to the folketymology here is in the thinking that this is the only meaning that can be obtained from this name. It is the meanings behind the names of the characters in the myth that provide the backdrop for the narrative.” Matthew 1:21. because Ogún (Qayin. not Yah who judges (or saves).” In other words. they do not treat him like one would do a normal character in a myth. war. rid. escape. Throughout the narrative(s) each trait is highlighted in some extreme way to reinforce the definitions behind the character. based on Jesus’ own words in the Bible. For example. pluck. for he shall save his people from their sins. Cain) represents the spirit of iron. The word shua in Hebrew can either mean “save” or “wealth. We can contend that the Ya is just part of the word Yshuwa “save” which would make Jesus a “savior. “Moreover. deliver (self). recover. Because this name is not native to Hebrews. It is this scripture that can pose some problems in the meaning of the name.” But the scripture doesn’t speak about Yahweh saving. the Father judges no one. Proponents might argue that Jesus is God and therefore this is appropriate. and you shall call his name Jesus. preserve. Often these traits are embedded in the host language in a series of lexical items that share the same consonantal root system. to extricate:--deliver. take (out). to snatch away. it is Jesus. the Hebrew’s folk-etymology came into play as they attempted break down a term that cannot be broken down in Hebrew. It is his mediator who does the judging (and saving) in Africa. for reasons that will be further explained later.SH5338 5338 ntsal nets-al' (Aramaic) corresponding to 5337. and within this g-n root is a word that means “to hunt” or “war. Because Christian practitioners assume Jesus was a real person. whether in a good or a bad sense:.X at all.” Historians assume that what is meant in the root of the name Yehushua/Yeshua is “save” based on the following passage: “And she shall bear a son. The Bible is no different for in John 5:22 it states. rescue. The Trinity debate is a separate issue that space will not allow us to get into here.” it is guaranteed that some myths will exploit this characteristic and associate Ogún with “war” or “hunting. 45 . spoil. defend. it is guaranteed that somehow iron will play a central role in the narrative in Yorùbá mythology. save.” Within this same g-n root system are words that mean “violence” and this is why Qayin in the Biblical myth “kills” his brother Hebel because Qayin represents (literally) “violence. The question here is whether the Ya in Yeshua derives from Ya/Yah/Yahwe or is simply part of the word Yasha. For people state the name means “Yahweh Saves. Remember that the Hebrews didn’t write out their vowels. As the great Egyptologist Serge Sauneron notes in his work The Priest of Ancient Egypt (2000:125-127): The Egyptians never considered their language – that corresponding to the hieroglyphs – as a social tool. 10:21. 24. We do not know just what his name meant. “the Hebrews” (Gen. trespass. a subtle connection that priestly erudition would have to define… This practice can seem childish and anything but serious. Yet its logic emerges if we try to understand the value the Egyptians placed on the pronunciation of words. Ogún). etc. It is no coincidence then that Hebel was a shepherd. In the Bible. passer-by. FULA in Senegal/Gambia. 29 46 . EBER (b-r) is the name of the ancestor of the ben-ey ‘Eber’ “sons of Eber.29 What we just mentioned was a major convention in ancient Egypt. pass on. European to exceed bridge. YaBal was the ancestor of those who dwell in ‘ohel “tents” and own mi-qeneh “cattle” (Gen. PEUL in Guinea. essentially the very means of defining the nature of the deities. It thus became a general practice. to show and reveal himself.” These are the ‘iber-iym. calling him Amun. This name for Eber is comparable to the wandering Fulani of west Africa known as ABORE in Nigeria. Any superficial resemblance between two words was understood as conveying a direct connection between the two entities invoked. and to find in it an explanation of the god’s name: “thus addressing the primordial god…as an invisible and hidden being. employed in all periods and in all areas of inquiry. PULO in Senegal/Gambia (Ful-be plural). In Genesis HeBel (h-b-l) was a ro-eh so’n “a feeder of sheep and goats” (Gen. The moment one understands that words are intimately linked to the essences of the beings or objects they indicate. alien. resemblances between words cannot be fortuitous. Observe: Abar (Hebrew): abo’r suwfah: boro (Twi): purboro-fo: o (Bachama): afara (Yorùbá): cross over. (2011). This was the case with Amun.” It is this same g-n root that we get the term “gun” from: a tool used to commit violence. the great patron of Thebes. a cosmic force. and in priestly lore it was the basic technique for explaining proper nouns. 4:2). Thus study of this language enabled them to “explain” the cosmos. ford See my upcoming publication Ogun. 11:14-17). The word HeBel (Abel) belongs to an ancient b-r root. 4:20). for them.hunting. It was word-play that served as the means of making these explanations. and BORORO in Chad. The b-r/p-l/p-r/f-l root can be found in many African languages. MOCHAVersity Press. they invite him and exhort.” and the scribes played on this resemblance to define Amun as the great god who hid his real appearance from his children…The mere similarity of the sounds of the two words was enough to arouse a suspicion on the part of the priests that there was some close relationship between them. but it was pronounced like another word meaning “to be hidden. African Fire Philosophy and the Meaning of KMT.” (emphasis mine) This was no different among the Hebrews whose mythological characters served many functions based on the consonantal root system as we have already seen with the name Qayin (Cain. it always remained a resonant echo of the vital energy that had brought the universe to light. pass by wind whistling by to trespass. they express a natural relationship. the consonantal root of the name shua’ embeds all of the epithets associated with Yeshua. pass by” in Judges 19:12b. The call of the dam-iym of your ehi is shouting to me from ‘adam-ah. they had to create a scenario for which farming could be introduced into history. A systematic comparative analysis demonstrates that many of the characteristics of Yeshua in the myth was based on the meanings associated with the s-ʔ or s-‘ root (/s/ and the glottal stop /’/). the Hebrews decided that at every chance they can get. she will no longer give you its vitality…[Gen. the language by this time had not been spoken for over a century before their endeavor. It is the “meaning” behind the names that tell the story. but between two populations of people who had different occupations. It is fairly easy for us here because ancient Hebrew. Now. but the consonantal roots. This is why we rely on the tools available to us in historical comparative linguistics to get to the more likely pronunciations and meanings of terms that are shared throughout related languages. Because of the historical ‘beef’ between pastoralists and farmers. 19:12] So based on the “meaning” of the names.” Therefore to reinforce this notion of farming as slavery. This was not a battle between two persons. which on the continent of Africa is s-l. and to create a starting point for which the occupation could be vilified. If you till the ‘adam-ah laboriously. more resistant to change over time than vowels are and are more reliable when doing comparative study. they scripturally placed a curse on all of the land for which the farmers labored. did not write out their vowels. Consonants are. Over time sounds change in all languages and diachronic linguistics seeks to understand these series of changes over a 47 . We will ‘aBar until Gibe’ah [Judg. The Masorites and Tiberians were responsible for many of the alterations of vowels and definitions of Hebrew words during the 6th to 12th centuries CE. We shall not turn aside into a city of strangers who are not part of ben-ey yisera’el. Because the Hebels did not like the Qayins. they will take a shot at the occupation of farming. they were always seen “passing by” the cities of Canaan. This will become clearer as we move forward. we can deduce that the Cain and Abel story is myth about Farmers vs. morphologically. with the # sign representing zero or no sound value (the /s/ can also be a /z/ or /j/ value). 4:10-12] With this said. The b-r root means “pass on. them being nomads. accursed are you from the ‘adam-ah which opened her mouth with an oath to receive the damiym of your ehi from your hand. Pastoralists. It wasn’t until 1948 that it was brought back to life again after 1600 years. s-r or s-#. Sound Changes What makes the issue of pronunciation so difficult for ancient Hebrew is that the language had been dead or dying for a very long time. short cut (across an area) breeze The Hebrews got their name because. No. 18a and in Ruth 4:1.ibaibara: ra-(mu): afeeburu: re: across (the nose). In that. in mass: never settling. The Iber’iym idolized the pastoral lifestyle and considered farming “slavery. just like in Ancient Egyptian. This accounts for why many word-meanings are unknown to modern scholars and why we cannot solely rely on Hebrew texts for clarity. for example. For example. Because of this. One examines the cognate (matching) forms both formally and functionally.. Sometimes a reed-leaf (j/i) was used in place of the original r. the r consonant tended to disappear at the end of words. but it can also be left out: for example. but we pronounce it as if it had the final r: Wsr. then the latter represents the newer function (Childs..e.e. it has lost phonetic substance—like what can be seen in the pattern of English gonna from an earlier going to—the fuller form is likely the earlier one. 30 48 . from a lexical orientation to a more grammatical one. but others can omit it: for example. sometimes the scribe would use a spelling that more closely reflected the pronunciation of Middle Egyptian. This is what I think happened with the final –r in the name Wsr.” The word ntu means “to have being. to have. The Niger-Congo form is the older form because met’ey is unanalysable in Hebrew. The following consonants tended to be the most affected: r. More conservative spellings still show the r sign. man”. and its child branches that derived from it. nbdt “braid.e.” Igbo mmandu “person. See Janheinz Jahn’s seminal work rightly titled MUNTU: African Culture and the Western World. If the form has undergone semantic expansion and or shift. For instance. the Hebrew word met’ey “person of” is cognate with Common Bantu muntu “person. Egyptian mt “male.. even if one of the consonant spellings had changed over time. feminine words usually ended in t. If the form has suffered phonetic erosion.” Sumerian nitah. 1865-1870 CE). a series of patterns are recognizable and it helps us to understand what has changed since the beginning of the pre-dialectal parent language. i. Once in a while the scribe combined the “traditional” and “modern” spellings by showing both the r and j (reed-leaf) graphemes (Allen 2010: 20). and which is the innovative pattern in such an analysis is usually readily apparent. lord. eṣu. spirit. owner. with mu.span of time (i. this consonant had probably disappeared at the end of words. to have life” (a thing). You will see in later periods that the name is spelled sometimes with a final r and sometimes not. the classification of Muntu isn’t relegated to physical human beings as it can apply to God and the ancestors. 1994. 2008: 156).” The full form of the word is mu-ni-tu. Establishing which is the more dominant and/or older pattern between the compared languages.” Wolof nit “man. man”. Grove Press. Egyptian Hieroglyphic writing was conservative.. Conservative spellings still show it.” The word tu (ti) is the root and generally refers to an existing “thing or being. that is. The word ni. D. Ngome motu “man. many terms written in the Middle Kingdom period and beyond still maintained spellings from the Old Kingdom period. particularly in ways such as those discussed in Heine et al. T.means “person. and l. Therefore. possessor of. 1000-2011 CE). 2010. versus a synchronic analysis which seeks to understand a language at a smaller interval of time (i. Also see Placide Temples Bantu Philosophy.being a Niger-Congo noun class prefix (usually denoting intelligence and why it is relegated to the human noun class).” also spelled nbd(t) (Allen 2010: 20) However. We also note cognate terms in Yorùbá that have lost the final r sound as well: i. dAr “subdue. Throughout such an analysis.” which is written dAr or dA.e. ni “man. the m in Met’ey is a lexicalized mprefix from the Kongo-Saharan languages. HBC Publishing. A similar process helps us to detect erosion in languages and it is by this same process that we understand an erosion happened in Hebrew with the s-r/s-l root which turned the r/l in Wsr into a glottal stop in the name Yeshua’. In Egyptian. However. 1993.30 The Hebrew language does not have a noun class system like that of Niger-Congo. By the time of Middle Egyptian. So the word muntu is an “intelligent being” (a human being). t. We have a doublet in Egyptian. Koma gil “see”.. I base this on the following: PCS *(kV)mV “eye”. riran “to see” glyph. Mande n’ya’kise “eye”.” with the latter more likely being a loan. the glottal 49 . 'n. The Glottal Stop From here we can see that the initial k/g sound found in the Kongo-Saharan languages became a glottal stop (‘) in the Semitic languages (k-l. The glottal stop may derive from a host of different consonants. Hebrew and Egyptian lost the initial k/g value. In Sumerian old prefixes tend to be reduced to meaningless vowels. We find this same activity in Hebrew as will be shown further below. inu/enu. etc. Atlantic : Djola. “regarder”.t “eye” :ĕyĕr.sound became j and final r/l was dropped. Arab. Phoen. 'ayna.The letter r in Egyptian writing is represented by the “eye” r sound is the initial and dominant consonant of the word. in other words it is a sound that distinguishes one word from another. Pharonic Egyptian Coptic Yorùbá :ir. Akk. but an allophone i. Ugar. gi (250). ena or 'ayna. kil. 'ayn. ye-li ke. “eye” :ri.e. It takes the r value because the It is my contention that the Egyptian form is a variation of a more ancient variety that had an initial consonant before the r.’-yā-kili “eyeball”. for it has both of the latter forms of this word: ir and an “eye. 273b). We regard the glottal stop onset of German words with initial vowel as further evidence for NigerCongo CV prefixes in Indo-European. 273. In discussing the Lost Laryngeal. The glottal stop is produced when the vocal folds are compressed. The glottal stop is a phoneme in Arabic. (attested in Hittite as h. OSA 'yn. That means that pronouncing the word water with a glottal stop i. In African-American Vernacular English you will often hear people pronounce the word heaven as hea’em: replacing the /v/ with a glottal stop and the /n/ value turning into a /m/ sound. 'ayn.e.. Syr. şil “eye” (270. weakens and is attached to the preceding CV root. wa'er does not change the meaning of the word in English as it would in Arabic. or pinched together. The l sound morphed into the n sound in many of the Afro-Asiatic languages for which Hebrew belongs: Egyptian an “eye”. but generally lost in other Indo-European languages). Mande ye “to see”. It appears there was an initial k or g sound before the final -l. Buduma yil “eye”. both to produce sounds by sucking air in (implosives) and by pushing air out (ejectives). he states the following: In his introduction Chantraine briefly alludes to Indo-European “laryngeals”. Prefixes and suffixes alike tend to be eroded and lost due to weak accentuation. Jewish Aram. which is to be expected in Yorùbá. The word for “eyeball” is eyin oju which appears to carry both forms of the word: ‘ayin and ji. old suffixes tend to be reduced to a consonant. a variant of the phoneme /t/ in English's case. In the production of these sounds the glottis is used as a piston. Nilo-Saharan : Fur agil. A-A Longarim (Chari-Nile) gini. The glottal stop reflects a reduced consonant. GJK Campbell-Dunn in his Sumerian Dictionary (2009: 8) further enlightens us on possible developments of the glottal stop in the regions of the Middle East. The Yorùbá word for “eye” is oju. The g.'n. Eth. The ir form was the most dominant form in Ancient Egyptian. which becomes meaningless. n. a variant of a sound. ‘-n). and then released as pressure builds up. Pepel. “Bantu” ji “eye” (249). k-n > ‘-l. which we have explained elsewhere (Campbell-Dunn 2008) as remnants of old Niger-Congo CV prefixes. As Allen (2010: 15) noted in regards to Egyptian. gi. It isn't a phoneme in English. yĕr. (per Westermann) -kil “eye”. 32 which itself often derives from k/kh. the Khoisan h of Hazda etc. 2009b: 8) Many publications speak about how in different languages the glottal stop replaces the k sound.” Ayele Bekerie in his book Ethiopic. CN Moru kuru “head”. and in Middle Egyptian it may still have had that sound in some words in some dialects.25 31 50 . I believe that it derived from an initial k/g value with the y serving as a semi-vowel. p.147 32 See Murray B. Afro-Asiatic : Semitic : Aramaic gulgulto. kuru “head”. which ultimately became our English letter A (through the Greek alpha). See Lehmann (1955). where ha = *ka. This discussion on the glottal stop is important for us here because it helps to support our notion that Wsr/Jsr and Yeshua/Yehoshua are in fact the same name. Mande ku “head” (lost prefixes affect the initial consonant: T/K). In Hebrew. So it must be a weak fricative.stop ‘ may have sounded like English d as in deed. Motilal Banarsidass Publications. p. dama “ox”. head. The ancestral forms of this sign are as follows: Modern Hebrew aleph Rashi Phoenician PaleoHebrew Aramaic א However. leaving a bare initial vowel (ara “road”). Laryngeal Theory grew out of the work by De Saussure (1887) on vocalic coefficients. 1994. like the various Indo-European “laryngeals”. but lost in most IE languages. He informs us that: The Hittite “laryngeal” h is traceable to weakened Niger-Congo consonants (often from a prefix). this form is older than Semitic and derives from the word #kV “cow. as in hada 2 “brilliant”. John Benjamins Publishing Company. Keiler (1970) etc for the IE details. for the glottal stop at the end of Yeshua’ (‘ayin. In regards to ‘ayn. Compare the Niger-Congo prefixes in h. /ʔ/. Sumerian gud. Sumerian kir “cow” (fem). The evolution can be seen in the following image: See for instance KM Petyt. Sumerian h is from PNC k. The aleph אsign (the first sound in the Hebrew alphabet) is used as a consonant in initial and non initial word positions and is represented by a glottal stop /’/. It is a remnant. provides us with a historical context for the evolution of the term that became the Hebrew aleph. Often it can be omitted. ah) corresponds to the r value at the end of Wsr/Jsr. or it can be silent. an African Writing System (1997). it is less common than the replacement of d or t with a glottal stop in this dialect. Mangbetu ga “horn”. The Yorkshire dialect of English in Northern England is such an example. gu4 “ox. In many ways it resembles Hittite h . However. Emeneau Dravidian Studies: Selected Papers. 1985. Mande gama.31 The Dravidian language glottal stop in initial position may derive from an initial h. Berta alu. “skull” (cf Golgotha). (Campbell-Dunn. ha-ra-an “road”. ox” (ga “horns”). Dialect & Accent in Industrial West Yorkshire. Campbell-Dunn continues to add insight into this phenomena in Indo-European which we see happening in Semitic. In the Ancient Egyptian writing script this sign had the sound value of ka. power. The Egyptians had another way of expressing this term ka by using either a pair of raised arms or a full human body with arms raised imitating a cow or bull’s horn. to end”). the difference being the Greek Pi has two “legs. the Greek system followed the Hä (Alpha)/Pä (Pi) order. So in writing you will often see this sound written as ‘ or ʔ which some people articulate as ’ah (<ka) One significant parallel between the Ethiopic script and the Ancient Egyptian writing script is with the first and the last syllographs of the Ethiopic: that is (Hä /Hoi) and (Pä) [The syllabic order]. PWN KAKA “grandfather. master”). “elder. which matches conceptually with the Ancient Egyptian. priests. The word ka in African languages can mean. Pä means the end and is represented by the tail end of the lion (Egyptian ph “to attain to.The letter A in English is simply an upside-down set of cow horns. So the development of our English sign for A comes directly from an African sign that had an initial k.value that was lost which only left the vowel (a). king. kaka “chief. head. ancestor and chief. However. The horns and the hands are pointing upward which is signifying something “on high” or of “high rank” or “status” (the Creator. force. But for a time.is not fully lost. in Hebrew the k. PWS ka . LuvaleBantu Kaka “God”. king” (head of the tribe). from ro (ρ) to omega (Ω) to complete the series. The Greek pi even looks like the Ethiopic Pä. it is the beginning of all things and is why it is also associated with “God” (Egyptian Kaka “god”.” 33 51 . The Greeks added 8 more. the highest”) and is represented by the front end of a lion. it is simply weakened into a glottal stop. from alpha (α) to pi (π) come out of Africa.” With ka being the head. Hä in the Ancient Egyptian language means the beginning (ka “head.33 It should be noted that the first 16 signs of the Greek alphabet. grandmother. elders). blessing and celebration (Bekerie. which invokes glory.” There appears to be two major pictographic sources for Hä in Ethiopic: the raising of two hands by a standing person towards the heavens.htm for more details. The meaning has philosophical and theological underpinnings. call their sacred writing Kpa34 which is a match for HäPä. Among the Ga-Adangbe people of Ghana. whose oral history states they came from Ethiopia.com/GA-ADANGBES%20AUTHENTIC%20ANTHROPOLOGY. The first graph (Hä) is named Hoi. Ka. praise. The image below of the Ethiopic symbol for Hä better demonstrates this point. The term is also linked Visit. Hä) is the first letter of the alphabet.HäPä among the Ethiopians describes the writing system (Fidel) as a whole from beginning (Hä) to end (Pä). It literally means “the beginning. 1997: 84). and the frontal view of a long-horned ox or cow which are so fundamental and so central to the lives and cultures of the region. 34 52 . It is no coincidence that the letter A (Aleph. i. 35 Keep in mind that r/l/d is a common sound shift. that poignantly captures the most salient expression of them all: “Ethiopia stretches her hands unto God.e. a chandelier:--bright. heleh “fat”. kàl “charcoal” PWS la. k/g > ‘ sound mutation:35 Africana k-l PWS kà. Soko hahala. “lightning” Mande M gã-ndi. light. into the glottal stop in Hebrew. 1) (also ma-‘owr a luminous body or luminary. kau “fire. kàn. kã-ndi “fire” Hebrew ‘-r/’-l ‘owr “lightning” (light in Gen. heleh hittah “cream of wheat. i. as stated elsewhere. Mande gā-di.with creation and creativity (thus why it is at the beginning). hearth” Mande kami “charcoal”. halah “milk”. Hä is also a proverbial symbol.cheerfulness. The cosmogonic outreach associated with the syllograph is significant. brightness. kālo “moon” Mangbetu kago “fire” Mangbetu eka “to light”.) hlh “white”. kā-ndi “fire” Mande ta “fire”. ko “sun” Bantu jadi.” The following is a chart of the evolution of some of the Ethiopic characters and their ideographic meanings: The following table will compare some terms for which we can demonstrate the weakening of the k/g value in African languages. specifically. (dā) “day” PWS tá “fire” PWN KÁN “daylight” PWN ka “home” (home fires) Bantu káda “charcoal” Bantu kanya “to shine” Kongo ekala. Swahili makaa “charcoal” “Bantu” (Johnston 1922 : 296) kω.e. (abstractly) light (as an element): figuratively. Yorùbá èluhó iṣu (yam) “flour”). 53 . Hä is an abstraction of stretched hands – hands that are stretched towards the heavens.” (Zulu hlope “white”. sky. Bari ko. Jolaos (j-l) “sky. sky God.l “sun”. 273. a god (Horus)” Canaanitic El / B-el “God” cf. Buduma yil “eye” Atlantic : Djola. lord. şil “eye” (270. day”. fireplace” k-l/g-l PCS *(kV)mV “eye” PWS nì. Mw-ene “top. 273b). K-ulu. n. nìa “to see. Cushitic : Quara (C) yil “eye”. *kiki “eye”.ŋ. heaven” Egyptian hrw “upper side. si-nòŕo “sleep” Mande n’ya’kise “eye” Mande ye “to see”.e. i. Bangi. G-ulu. Poto lokiki “eyebrow” Ngombe etuki “eyelid” Bangi. Judging from PWS kia “become visible” the word “see” was once voiceless (no passive). (so Westermann) have -kil “eye” k-l/g-l/n-n Mande Nala “God” Bantu Kala. was a noun. Bagirmi kada “sun. Sardinian /Aegean Julus. Pala/H-ala. possessor. (passim) & işi. “regarder”. Pepel. kè. ye-li ke. eye” PWN NÍN. (n)K-ale. ye-li “eye” Mande n’-ya-ni “eyeball’. Turkana ekoloŋ “sun” (Greenberg) Tamil culli “furnace.lo. K-ule. Ngala. i. ix. Qara yil “eye” A-A Longarim (Chari-Nile)gini. ñωhũ “eye” (260) Common Bantu (Stewart) ki “to dawn” “Bantu” ji “eye” (249). nū “eye” (258). Z-ulu. Lolo.. top. ihi. nù. “sky. kíu “moon”. owner” Akan Osoro “a sky God” (see Darkwah 2005) Amazulu zulu “sky. Z-eru. Mande ye. kil. ‘eloah “God”(‘lh). heaven” (God?) (with masculine suffix –s) ‘-n (l>n) ‘ayin “eye” Afro-Asiatic : Chad : Buduma (2) yil. Arabic ‘allah 54 . Semitic : Arabic ?inna “behold”.’-yā-kili “eyeball” Mangbetu ki “visit” “Bantu” (Johnston) numì “eye” (257). etc (Johnston 1922 : 290) all suggest early *gigi. Mangbutu kora “day” CN Didinga kor “sun”. Jolos. Dinka ako.CS Kreish kadda. kìa “become visible” may be connected (first syllable). iş. NÍNU “eye” CS : Madi mi “eye” PWS kí. isi. God” Bw-ena. surface. Bamana ŋa “eyes”. Hebrew hinn“behold” ‘-l ‘el. ih’. G-udu. gi (250). Koma gil “see” Chari-Nile: Sokoro yi:di “eye”. Ngala keka “look” Nilo-Saharan : Fur agil. This ‘ayin ַ ֻׁ ַ character corresponds to the Egyptian sign for the eye: ir . 1995: 36) 55 . It is apparent that the Greeks utilized the Aramaic version of Yeshua (j-s-‘) as they pronounced it Iesou-s with masculine –s suffix (j-s). When we put the consonantal skeletons together. j-s-r j-s-r s-r / j-s-r j-s-r In the Greek rendering of the name Wsr. but it is noteworthy that Plutarch. It’s important here for the fact that the Hebrew name Yehoshua י ְהֹוׁשעends with the ‘ayin עcharacter. De Iside et Osiride. The j. Here the prefix wa.prefix. 36 37 Usually a consonantal y. oyoser – ouoser). Yeve. Remember that the y can also be an i.or w. at the beginning of words sometimes identical with A (Gardiner. Wsr in Coptic is oysire (Ousire).became an o.sounds in Wsr are just prefixes or vowels. Jsr “Asar” Ye-shua’ j-s-r j-š-‘ Yehoshua yh-š -‘ The orthographic j can be either an a. e and i y i y 36 i37 y j (yod)later > /ʔ/ jj or y j or i or yod y With this said.We are discussing this at length here because the Hebrew aleph and ayin sounds are the same in common usage in Israel and it is my contention that they came about by the same process. In Coptic we see find the /o/ sound as well: Middle Egyptian Wsr “rudder” (Coptic oyosr – ouosr. Osiris (Osiris) in Greek has a different first vowel. Many words in Hebrew with the glottal stop correspond to other African terms that have a k or g value. In the ancient Egyptian. It should be reiterated that the glottal stop can derive from pretty much any consonant. which Hellanicus of Lesbos. we can see visually how the name Jesus/Iesous/Yeshua was written in Egypt before there was a New Testament or Hebrew people (as Asar is attested from the 3rd or 5th dynasty). i or y. The root of both Wsr and Yeshua is s-r/š -‘. Budge (1920) Gardiner (1957) Mi Ra (1995) Loprieno (1995) Allen (2010) a. 34.sound in Greek. Another form shows another prefix t-shuw`ah. 1957:27) Long ē as in English “easy” (Mi Ra. it begins to make more sense. heard the priests pronounce. The form yasha/yesha is just a variant as we will see further below. Hebrew did not carry the w. the reed flowering leaf is given as the sign for the sound j or the Hebrew yod. according to him. Ya) or it could simply derive from the root yshuwa as demonstrated in the Strong dictionary (#3444). The j/y in Jesus can derive as a shortening of Yahweh (Yahwe. 364 D records a form usirij (usiris). they pronounced it Osiris adding the suffixal –s. This helps us further demonstrate that the root of Yeshua’ is š-‘. The Egyptian Book of the Dead. the eternally good being. request. lightining” sola = clear. esteem. fancy. savior. great. gba “obtain” toro “beg” (ts>t ?) [fe “love. obtain. love. Almighty. desire” sAir “need. king. want. day. explicit See the Wikipedia entry for more commentary on Yeshua’s epithets: to the ROOT of the Problem We mentioned earlier that in myths the main characteristics of the main characters are often embedded in the root of the name. beg. paraclete. ni. shepherd. anointed. tough” sAr “wish. Yorùbá and ciLuba languages to note the many sound changes. new Adam. prophet. white light” (often given as “white cloth”) Swila = love. mighty.wikipedia. anything strong. wish”) Kikongo zola “love” Wsr “to dry up” (with sun glyph determinative) ‘ahal “clear. "the one who governs Ro-Setau". The following table will demonstrate many terms that belong to the s-r/s-l root that match many of the epithets given to Jesus the Christ and Wsr (some yet to be mentioned) utilizing the Ancient Egyptian. desire” sAry “needy man” ri-gba. power" 'el-ah "an oak or other strong tree" (feminine form) Sha’al “desire. head of the church. creator of millions. 38 56 . ife “wish”] [Igbo yi oyi “friends. son of man. many-faced. good shepherd". clear. strength" 'ayil "strength."38 Wsr "the great inert". be strong. power” Hebrew uzai/uwzay "strong" 'uwl "twist. desire. shine” ‘owr “bright. specifically a chief. lord of lords. and life. also a ram" 'eyal "strength" (a variation of SH353) 'el "strength. n’anya “desire. 'the way'. love Yorùbá le “strong” (with dropping of initial s-). wish”] ala “white. judge of the blessed dead. "the first of the westerners. lord. desire freely (<swa “love. truth. desire.org/wiki/Names_and_titles_of_Jesus_in_the_New_Testament 39 See Raymond O. the perfect one. Faulkner. mediator. prefer. like. apostle. The following table displays many of the more common epithets or titles (characteristics) of Yeshua and Wsr of ancient times. logos.39 It will be demonstrated that a number of these epithets are associated with our s-r root. "the Lord of the living". Terms associated with the s-r root Ancient Egyptian Wsr “strong. deliverer. nipa-lera “strong” (most common form lagbara) ciLuba shili = strong. mighty. the lamb. demand” Ra’yah (feminine) a female associate. Jesus "priest. solid -shilè(à) “hard. fellow. begetter. "the king of those who are not". Hebrew. leave. fain. strength. lovers”. care for. "the master of eternity". speak with intelligence sare "to make right. in perfect order. Lolo. We also note from strong’s dictionary: Gi-ychown ghee-khone' or (shortened) Gichown {ghee-khone'}. san "aloud. Lolo basi “water” etc]. song or the like). make an offering (of food or drink)” ‘olah “burnt offering. true" muSulu = stream. to be right" sòre "to do good" san "aloud. to say” (s<kh by way of palatalization?) san “beautify” (r>n) swnwn “to speak flatteringly” m sSr “all right. ascend” (feminine active particle) sàra. intransitively. to send up. erectly" shili = be civilized bu-shuwa “truth. compare. luselu “all goods that a man gives a woman to marry” The pi. adage. complete. Mangbetu gi “river bank”.xr “speak. to appear” ‘ala-h “to arise. fulfill (a contract)” usn “to make water” Piyshown “a river of Eden”40 sare "to make right. set right (a wrong). Poto mai “water”. finish off” sar “bring to an end. as a stream. artery. correct. directly. provide. i. sàráhà "alms. Kele wato “boat” (bw > Sumerian m ). also a valley (or pool) near Jerusalem:--Gihon. straightforward. directly.” wi. finished. etc. muShilu = vein. raise. river MuZilu. lift up. to be right". excel. stream. Ngombe bwatu “boat”. utter. say intelligently. Poto. completed. straightforward.” saŋ "uprightly. vividly" zara-kh “to rise. But Bangi. to make rise” sary “stand” (for container) Jula / Shula = to lift. use (as a) proverb. as thunder" Oṣun “river goddess. to burst. watu “boat”. make restitution” sar “present. a river of Paradise. approach death” Shalam “perfect. so. arise. in good condition” Ma-shal “to liken. finish Shila = be reduced to ashes. charitable gift" (may be a loan) sula = be at the end. wSn “wring a neck. (a path) Sark “complete. Soko. from 1518.e. River Oṣun” Jala = “quite. become Dikala be carbonized sela “to pay a dowry for a wife” (also to marry a woman). raise. Ngala. to resemble:--be(-come) like. finish. vividly" saŋ "to be loose. soro “speak” Zela = display. to flow. ascend. to sacrifice. fetch up. (transitively) to use figurative language (an allegory. speak (in proverbs). freshly straight” bushuwa/bushiwa "reality. Ngala.may be an old prefix: Bamana ba “river”. carry up. reality” (<-lelèlè(à)) sar “cause to ascend. raise up srwD “make secure. to publish. to burst forth. finished Shila = complete. Bangi bwato “boat”. cast up. Notice the Mangbetu term for “river bank” gi prefixed to Gi-ychown? 40 57 . sacrifice. Gichon. fleeing from danger.” I did however find jilii (j-l) “to remove” and hallas “rescue. to flee" Sálo "to run away.” The s-r root is reduced in the Yorùbá term olá "that which saves. MuJilu The Yorùbá language appears to have more morphological variety than the other three languages compared. extricate. liberate. to run. a fruit resembling a w alnut" Asalù "having recourse to another for protection" Asáré "a runner.” To “save” or to “rescue” someone is to break them “free” from something that is hindering their movement or growth.” In the ancient Egyptian we have the (l>n) form in the word sni “to rescue. olà “wealth” (by way of initial consonant loss?) sera eni "to practice self denial" kale “rich”(s<k?) sAw “restrain” (A = l?) Jila / zila / Shila = abstain. protection. to cut out. gird.Wsr “wealthy. let go. undone. sure "to run. to secure. rich. Here in Yorùbá we see different variations of the same word with the final consonant mutating in the expected r/l/d (n) manner: Sádi "to take refuge under the protection of another" Sálà "to escape. pleasure. to rescue. Yorùbá and Tshiluba s-r root which belongs to a conceptual theme of “saving. These terms are in the Tshiluba language as shila “leave” = shiyila. to recover. to take away. All medical doctors. For instance. I couldn’t find a reference for the word shua’ to mean “save. They help people “flee” or “escape” from sickness and death: they protect from illness. since we are arguing that the s-‘ root in Hebrew corresponds to the s-r/s-l root in African languages. running." The later term is related to all of the other Yorùbá forms cited above. to maintain. also called Asúré" We have an agglutinated term in Yorùbá igba-sile which means “to rescue. and escape from a bad situation. wealth) ìṣura “treasure. Some messiahs help free people from slavery. tshua’ah) in the Yorùbá language. for example.” In Egyptian we have Sdi "to break. yshuwa’. salvation. join together. elope" Sálù “to beg help of another" sáré. Qe-sher “an unlawful alliance” 58 . asking for help. sacred> CiZila. to save. free.” In Yorùbá we have the word sádi “to take refuge under the protection of another. become loose. powerful. influential” Shuwa’ “wealth. delight” ṣùru "plenty" (abundance. liberate”. gallop" Àsálà "an escape.”41 In the Canaanite language. break its links. the cause of salvation. give freedom. Harriet Tubman is our most well known messiah among the Bakala (African-American) people. sula = break free. sulula “untie. flee. we would expect to find terms within this root field that is related to “salvation” (Hebrew yasha’. are “saviors” (or at least are supposed to be).” Other 41 The opposite in Hebrew: Qa-shar “to tie. The Hebrew shuwa’/shua’ corresponds to the Egyptian. anything valuable” òrò. as it has many versions of the same word with various sound mutations not apparent in the others. combine”. A redeemer or rescuer is someone that helps one to break free. revered. to serve a superior.” but is a “title” for someone who embodies ALL of the traits mentioned above (righteousness. perfection.” To be a shuwa’. or set of consonantal roots. This is why Jesus had to “rise” from the dead. adore. sacredness. sacrifice. plentitude. characters in myths represent many traits found in nature or in the social make-up of human beings." Yehoshua/Yeshua came to earth to correct the wrongs alleged by Adam (Romans 5:12-19): Yorùbá sare "to make right. 59 . perfection and strength). and to assume a role of privilege. love.themes associated with Yeshua that are embedded in this s-r root. abundance. In the myth he was bound to exert these traits as they are part of the linguistic root that makes up his name. lord over.” When an Onitsha person uses eze Ozo he is using a shorter form of eze kwana Ozo yi or eke eze kwana Ozo yi! This means “do not assume any role. complete. prevent from doing. The associated traits for the characters in the myth often belong to an “associative field” of meaning that is encapsulated by a particular.” Eze can also mean “to honor. In the Igbo language. to lift. etc. sacrifice." It is no coincidence that Jesus of the Bible embodied all of these characteristics. to raise up. abundance and strength. cease. to be right. privileges. death. love. offerings. demand debt due.” The Igbo word Ezè means “king. lead on the way. to participate. to fulfill (a debt). to rise. The two entities must share other traits and it is this collection of traits that will establish a case for the Hebrew borrowings from Egypt. holy. offerings. to call a blessing upon. or obligations that accompany ozo title yet!” The following chapter will begin a series of other associations not discussed previously. domineer. this root is reflected as ọzō “title of high degree conferring on the owner privileges and honour as a sacrosanct (sacred. sacredness. In the case of a Yeshua. one must be willing to bless others: Yorùbá sure "to bless.” This is another reason why he had to be associated with “fulfillment” of a prophecy and had to have a special “death” to pay a “debt” because the s-r root means “approach death.” Another variation in Yorùbá of this root can be seen in the term sin (r>n) "to accompany. plentitude. because the s-r root means “to ascend. perfect. as demonstrated in the table above. We are building the case that Yeshua and Washil (Wsr) are in fact the same deity with the same name. We now go on to connect various motifs to make a stronger case for borrowing as simply having the same name is not enough to support such a claim. raise cattle" To recap. chief. make restitution. So it is no surprise then that we find in Yorùbá Asalu "a title of honour among the Ogboni people" (Egyptian sr “great one. untouchable) being. are “righteousness. perfection. this is a term that doesn’t only mean “save. keep domestic animals. an official). but aspects of the same energy: the binary nature of life. This is why Oṣun gave birth to Eṣu because it is believed the feminine principle preceded the masculine principle. She is the personification of beauty and sexuality. out of his belly shall flow rivers of living water. The good ajogun control wealth. Another verse that supports this interpretation comes from John 7:38 (KJV): “He that believeth on me. We note also that Oṣun belongs to the s-r root discussed earlier where the r >n. They are the same energy expressed as “masculine” and “feminine” for the sake of the stories and to highlight certain functions that would be best demonstrated in this manner. like Jesus. What is going to be critical here for our comparative associations is that when the Egyptians wanted to emphasize a particular aspect of Wsr.” However. Osiris and Isis are not separate energies.” If we interpret Yah to be associated with “water.” We will also introduce in this section for comparison the Yorùbá òrìṣàs Eṣu and Oṣun. the one who carries the staff of god. 1995: 73). they would make it apparent in the iconography surrounding him in reliefs. The evil ajogun control death. Whosoever drinketh of this water shall thirst again: 14 But whosoever drinketh of the water that I shall give him shall never thirst. Wsr’s association is primarily in the physical sense. As we discussed earlier concerning the storm god Yahwe/Yahweh/Ya. 60 . in Egypt these concepts are seen as husband/wife AND brother/sister. love. love and rivers. To the thirsty I will give from the fountain of the water of life without payment. the trickster.” This can be supported in the Biblical texts as it is stated in John 4:13-14 as Jesus is talking to the woman at the well: Jesus answered and said unto her. "It is done! I am the Alpha and the Omega. This is a poetic way to say that the concepts are closely related. Again. and Shua to mean “life” (wealth. Eṣu rules through the ajogun. children. “And he said to me.” So Eṣu and Oṣun should not be seen as “separate” entities. the beginning and the end. the divine messenger.” We also have in Rev. the same concept is at play as these relations are to demonstrate the closeness of the concepts they represent. illness. Ultimately. What happens in Africa often is that they will “divide” the concept being personified into “masculine” and “feminine. and so on. through sacrifice. 21:5-7. as the scripture hath said. I think we have enough evidence here to support the notion that the name Jesus primarily means “Lord of life.” as discussed previously. serves as a messenger between humans and the other òrìṣà and between humans and God. (Neimark. Eṣu is the “owner of the crossroads. We now move on to discuss specific distinct traits for each of our “deities. success.” He is the force that “activates” or causes things to happen. Eṣu and Oṣun are different aspects of the same conceptual idea and it will become evident later on in our discourse. water is symbolic for “life” in practically all indigenous cultures. The goddess Oṣun is Eṣu’s mother. loss. or in the unique way in which his name was spelled. existence. In Ancient Egypt these two Yorùbá deities are Wsr and As." Wsr. What each of these names and spiritual forces have in common is the s/sh root that is related to “life” in all of its aspects.JESUS/OSIRIS “LORDS OF LIFE”…AND WATER! It has been established that the name Yeshua/Yehushuwa of the Hebrews is the exact name of Jsr/Wsr of the ancient Egyptians: just a dialectical variation. “blessings”) then Yashua/Yeshua simply means “living waters. She represents the generative life-force in the universe. wives/husbands. mental unrest and similar forces.t. is also associated with “living water. Eṣu. but the water that I shall give him shall be in him a well of water springing up into everlasting life. As Oṣun and Eṣu in Yorùbá have a mother/son relationship. linguistically. a wise priest as can be denoted by the “eyes” (consciousness) and the “throne” (seat of power.Ancient Egyptian wsr = Ras. of both Eṣu and àṣe belonging to the same s/s-r root that means “life”: the same root found in Wsr and Ast. God granted Eṣu the force to make all things happen and multiply (àṣẹ) (ibid. the word is to be rendered ras “head. Ashil. destiny. Olori “chief”). As Thompson informs us. iri) symbol with the r sound value is placed in front of the throne m (as) symbol with the s sound value. This is the period in which the following versions of the name of Wsr were primarily used. When trying to articulate this point. spirit forces” (<ori “head. This is a result. chief. king. However. Since Wsr is also the founder of Egypt in some stories. and it is my belief that many of the different ways of spelling the name are in fact different pronunciations that render different meanings as discussed in Imhotep (2011). there are dozens of ways in which this name is written. Thus. This became very pronounced during the Ptolemaic period when the Romans ruled Egypt. 1984: 28). this rendering can mean “primordial ancestor” or “founding ancestor” as all “kings” in Africa are just reincarnations of the founding “king/father/head” who founded the settlement. This life principle is reflected in our s-r root in Egyptian in the term siwr "make pregnant” (to cause life).” It is this root that is found in the Yorùbá word òrìṣà “head. Òrìṣà The above rendering is the most common way to write out the name of Wsr. But Wsr is associated with more than just administrative authority. God just means “ancestor” as all things derive from it. place for which one makes commands). This rendering informs us that Wsr is a king. “Eshu represents the principle of life and individuality who combines male and female valences” (Thompson. He is the very life-force in the Universe. these renderings were based on prior associations of the characteristics of Wsr: Different spellings of the name Wsr (Wa-shil. however. Eṣu among the Yorùbá is the principle of life and is often depicted as male and female in his iconography. 18). anscestors. In some myths he is the Creator himself. This is just symbolic of Wsr being our primary ancestor: the very energy of and principle of life. Mu-Jilu in Tshiluba) A 1 B 2 61 . the ancient Egyptians often employed an ideographic reading to the glyphs in his name. an elder. We will see the importance of this later on. consciousness. when the “eye” d (ir. The “acrophonic principle” during the Ptolemaic period is a form of figurative writing in which the name of a god is written with (and at the same time his theological qualities iconically evoked by) specific hieroglyphic signs used alphabetically (Loprieno. this may be an allusion to this fact as this area was called tA nTr “Gods land” which I argue simply means “land of our ancestors” as God (*godo. km = kun) the “spirit of the ocean” (kun “sea. Growing out of the pool of water is vegetation for which his four sons (also the sons of Hrw) are 62 . salt water. This whole scene can be read by way of either the hieroglyphs or ideographically by examining the poses of the figures. But what’s important for us here is not so much the judgment scene itself. the rectangular image is of a “pool” or “pond” and it normally has a /sh/ sound value. This is the symbol of royalty. In column 1A above. The trailing image is of a seated wise man which denotes people or spirits of high rank. but has entrusted all judgment to the Son [Jesus]. 1995: 24). pond. It was a way to fully exploit the “meanings” behind the sign.” It has been argued by many African authors that the origins of the Egyptians lay in the Great Lakes regions of east Africa. but what Wsr is sitting on at the far right. It is the first consonant of a pluriconsonantal sign that is kept. Wsr is the founder of Egyptian civilization and it is the founding ancestor who is deified and becomes the people’s totem in Africa. and all human life came out of the Great Lakes region. water. he has been vindicated of all unrighteousness and is now being led by Hrw to the inner sactum of Wsr (his higher self).” This is just one more solid correlation. This role is to later be assigned to Yeshua. After his heart has been weighted against the feather of truth (ma’at). or any deep waters). is sitting on a throne. there is a lake in Egypt called km wr and the lake is not black. “Moreover. The throne is sitting on top of the hieroglyph for “water” sw (siwa. as king.” One of his other names is km wr which many render as “the great Black. This image is telling us that Jsr/Wsr is associated with “water” and probably a “great lake. ziwa). As John 5:22 states. This demonstrates the role of Wsr as Judge in the afterlife. This lake or pond glyph carries the value of s or sh and derives from the words sa/za/siwa/ziwa/anza in Kongo-Saharan languages which means “lake. the Father judges no one. river.” However. Here the deceased scribe Hunefer is being led through the process of judgment starting on the far left. *gudu) is just a word for ancestor: so is nTr. Since Wsr is the “founding ancestor” of the people. From the Papyrus of Hunefer. regardless of the full phonetic value of the sign. I equate this to the Yorùbá word Olukun (wr = olu. and lordship in Egyptian iconography. Wsr. 19th Dynasty: Judgement scene The above relief is from the Papyrus Hunefer. ownership. or the iconography surrounding the figures. in part.” (Kiswahili siwa/ziwa “lake. pond”). Chief of the heaven! In thy name of Conductor of the Heaven. It derives from a verb Sänäyä and it means the “act of being learned and beautiful.jrank.”42 (emphasis mine) The symbol for the collection of reeds or other plants on top a body of water represents a variation of an s/sh sounds. The first is si (of the class sä ). Bekerie (1997: 120) discusses an Ethiopian philosophical idea called Se’en which is a term for “aesthetics. let the Osiris prevail over the waters…” Budge informs us (1899>1967: cxxxvi) that in the BD 57 that. 58 and 59 are titled chapters for “command of water” or “having power over water. Sound value for read and pool glyphs s s š 42 See. By studying the Ethiopian Fidel script.” Se’en in Ethiopic writing is composed of two syllographs.” With Wsr sitting on the throne. This root can be seen in words for breasts in many Bantu languages: Rukwangali-Bantu ma-sini “breasts.org/NUM_ORC/NYANZA_from_the_ancient_Bantu_r. and the form NYASA has become the proper name of a particular lake.” Setswana-Bantu ma-shi “breasts”. After death one becomes one with Wsr. This root can also be observed in the name of one of the major Great Lakes in Africa. “The recital of this Chapter gave the deceased dominion over the waters. a river or lake).” It is a collective term that represents the actual products of cultural undertakings.root. Nyanza: “NYANZA (From the ancient Bantu root anza. Wsr = “Lord of Water” As we stated earlier. The word is variously spelt. “Oh Hapi.html 63 . the scene is telling us that Wsr is the “Lord of water and vegetation” and by extension “Lord of Life.” This is further confirmed in the spells of the Book of the Dead. states. Nyanza is the spelling used in designating the great lakes which are the main reservoirs of the River Nile.seen standing at the top of the lotus plant. The root is sa or za which means “a collection of water”: Egyptian swsw “a body of water. we come to understand the true significance of this glyph used in the name of Wsr in pharaonic times. especially applied to the Great Lakes of East Africa. Nuer. Nara sa “milk. Wsr is associated with water and the life that is maintained as a result of its existence based upon an ancient -s. a modern simplified version of mdw nTr writing.” Keep in mind that all of the deceased latched on the name Wsr at the beginning of their names.” siw “a body of water. This glyph is a simplified form of the Egyptian collected reed glyph variations seen in the table below. What this aspect of the scene is trying to get us to understand is “what Wsr is lord of.” Proto-Taman *sun “milk” (Dinka ca. spells 57.” Spell 57. Shilluk cak “milk”). the Bantu name for any SHEET or stream of WATER of considerable SIZE. A matter of fact. rain. garden.f f dwAt This is Osiris. siwa. This same glyph is found in the ancient Egyptian. This same root is given for the word jsr “plant” [Wb I. The -s.” Below are some terms for which the -s.root signifies farming and agriculture and anything associated with it (i. force green fruit to ripen" (sinpon "to force green fruit to ripen by exclusion of air").t "a town in the Other World" (Budge 90a). Egyptian sr “to reach down”). Arabic and Hebrew writing with the s sound value. in this instance. siwH "inundate"). etc. rivers. Wsr is the source of the Nile. So not only is Wsr the water and vegetation. l’s and d’s interchange. TA “sky” PWS la. an ideographic as well as pictographic name that signifies farming and agriculture. garden with pool”. jewels"). "to settle at the bottom of a liquid. lake. Oṣun and Yeshua are associated with fish). Gio de “day” Sumerian šèĝ “rain” -ĝ 64 . osan "name of a tree and its fruit. his circuit is the Duat The “bed” or “seat” sign above the eye glyph derives from the Egyptian word jsrw "an article of furniture" (Yorùbá Òṣó "furniture. he is the actual soil (the “underworld”) for which life originates on earth. RAIN DA. neatness.root denotes a source that has life sustaining liquids (lakes. In the Egyptian literature.) and is why this same root is associated with vocabulary meaning “life. What’s important here is that jsr (Wsr). asha (contraction of Ashárà) "leaf or roll of tobacco" It is apparent that Wsr is spelled different ways and each term represents a different concept with a different pronunciation. sin "to bury. ziwa “water. river. in other words. lime tree and fruit. is associated with the spirit that makes things grow. osùn "a kind of herb". Mano de “day”.The name of this order or class of syllographs is called Sawat. We also note that there is a city in the spirit world with the same root: jsr. he is the actual location of the underworld/underground (Yorùbá isale “down below. down". One way to spell Wsr’s name is Jsr/Jsir as can be seen in the following: Jsr pw Sn(w). Keep in mind that the /t/ sound often morphs into /s/ (ts>s) and t’s. S 132]. pond”. jsr “tamarisk” [Wb I S 130] (which is a tree). decorations. This same j-s-r root is in another term jsrd "to make grow" (Budge 90a). The second syllograph is ni (<nä “order”). clouds. be concealed. What’s important for us here is that the sä ( ) glyph. This particular passage describes Wsr as the underworld (dwAt) himself. (da) “day”. as expressed in Ethiopic.root for water derives. pond.” sílè "to the ground. Egyptian S/Swt “pool.. to form grounds". lake. Bantu sa.e. habit of dress. breasts. It appears this -s. signifies farming and agriculture: two themes associated with Wsr who we saw earlier had the “sä” glyph composing his name. kind of fish" (remember Wsr. These terms are reflected in Yorùbá as Aṣán "a plain vegetable diet without sauce or meat". finery. Lolo basi “water” etc]. It should be noted that the Egyptian term iSrw “water meadows” has the same root value of the word Wsr: jsr. Ngala (2) še: “drink”. As we can see in 2A and 2B in the previous table above. rain. river MuZilu. tua “water”. River Oṣun” ciLuba muSulu = stream. from 1518. Gichon. Mangbetu gi “river bank”. Ngombe bwatu “boat”..mb “rain”. The following hymn from the Pyramid Texts associates Wsr with rain and water in general: The pi. waterplace” Bantu dé “sky. Soko. (a path) We have direct textual evidence that Wsr is associated with water.nde “clouds”. THUA “river. also ducks and a fish (the fish acting as the /s/ value) that evoke a relationship with plant and wild life that exist along lakes and rivers. donga “river”. Poto. Logone (2) se “drink” *T = š *A = e *G = g Wsr is associated with “water” and the various spellings of his name depict this association using other types of glyphs which demonstrate a natural relationship to water. du. “cloud ”.” These terms derive from the same or similar consonantal root cluster to “form” the composite diety Wsr/Jsir. Because Wsr is associated with water. Notice the Mangbetu term for “river bank” gi prefixed to Giychown? 43 65 . Guang n-tśu’ “water”. du. also a valley (or pool) near Jerusalem:--Gihon.PWS ta “sky. So one cannot view Wsr (Jesus) one dimensionally. stream. Ngala. to burst. artery. he is also associated with the channels for which water flows which are reflected (with a by-form) in the following languages as: Egyptian usn “to make water” Hebrew Piyshown “a river of Eden”43 Yoruba saŋ "to be loose. raincloud” PWN TU “cloud”. watu “boat”.may be an old prefix: Bamana ba “river”. to flow. cloud” Bantu tu. Wsr was a “god” of the Nile River and represented the annual flooding which watered the banks along this river for which the dark. Ngala. We also note from strong’s dictionary: Gi-ychown ghee-khone' or (shortened) Gichown {ghee-khone'}. because the collective terminology used to composite the deity makes him multi-dimensional: multi-layered. As is demonstrated here. a river of Paradise. as a stream. muShilu = vein. But Bangi. clouds” PWS tu. Lolo.-e “water” PWN BUDA “rain. to burst forth. as thunder" Oṣun “river goddess. there are depictions of reed leaves rising from a swamp. Kele wato “boat” (bw > Sumerian m ). Bangi bwato “boat”. to “river” Kele use “sky”. THU. Afema a-su. a similarly rendered term is layered with other similar sounding terms within the same conceptual field: the “root metaphor. fertile soil emerged which was utilized to plant crops. Ngombe buse “sky” Mande sã “rain” Mangbetu tu “pool in the forest” ? Mangbetu ro “sky” Afro-Asiatic: Chad: Hausa (1) ša: “drink”. Poto mai “water”. [ … ] your rear [ … ]. 2005: 302): RECITATION.Hymn 685 (Faulkner 1969) The waters of life which are in the sky come. 45 The [ … ] sign denote that the original inscription in this area is not clear due to age and damage. for the testimony of God is this. 44 66 . but with the water and with the blood. More confirmation that Wsr is associated with water comes from Spell 604 (Allen. 7For there are three that testify: 8the Spirit and the water and the blood. the Dual Ennead will shoulder him. the Dual Ennead will shoulder him. the great MooringPost has libated to you. To further confirm the association of Wsr and living waters. 10The one who believes in the Son of God has the testimony in himself. 45 The association with Jesus and water is strengthened when we examine the crucifixion in John 19:34: “But one of the soldiers with a spear pierced his side. and the Sun will give his arm for Pepi Neferkare toward the place in which the god is. that God has given us eternal life. the testimony of God is greater. as well as other epithets for which we will exploit later on in this text. but he who believes that Jesus is the Son of God? 6This is the One who came by water and blood. He brings the water of life. [Linear A has RA2 “water”]. ro “sky”. see James P. 5Who is the one who overcomes the world. one has to ask one’s self. which comes from Osiris: wash your arms. not with the water only. and the Sun will give his arm for Pepi Neferkare toward the place in which the god is. This is a clue that the crucifixion is not a literal one. Society of Biblical Literature. O. 9If we receive the testimony of men. [Isis has … to you]. the earth quakes at you before the god's birth. It is the Spirit who testifies. we turn to Spell 548 (Allen. the waters of life which are in the earth come. Allen (2005: 291). Osiris will bear him up.44 Wsr is not associated with rain randomly. Yorùbá sèrì "to drop dew". because he has not believed in the testimony that God has given concerning His Son. The Ancient Egyptian Pyramid Texts. Nephthys has screamed for you. 11And the testimony is this. You have your water. and forthwith came there out blood and water.” Now. and the three are in agreement. Geb will bear him up. as we reiterate continuously here. Nephthys is opening [ … ]. When Pepi Neferkare goes down into earth. open up [your] ears [ … ]. old man (Pepi Neferkare) [ … ]. Clearly the earth quakes at the human birth of Osiris. 12He who has the Son has the life. 2005: 298): When Pepi Neferkare goes down into water. striking [ … ] and they will give you to [ … ] your nurturer [ … ]. what is meant by “water” rushing out of the side of Jesus after he was stabbed in the side with a spear? No human being has ever been stabbed and water rushed from their body: at least clearly where one could distinquish it from the blood also coming out. Jesus Christ. Nilo-Saharan: Fur ara “rain”. the root for his name means “rain”: Mangbetu zoro “rain”. For a different variation. [ … ] has bowed over his brother. We have in 1 John 5:5-12 another testament to the association of Jesus with water. that He has testified concerning His Son. he who does not have the Son of God does not have the life. This is because. because the Spirit is the truth. the sky is aflame for you. and this life is in His Son. the one who does not believe God has made Him a liar. Webster's Bible Translation And killed the Prince of life. the power to cause to be. then they are in fact the possessors or “lords of life. there is no life. Given these two things. Yeshua is seen as the generator and officiator of life. whom God gave back from the dead. Compare with Westermann’s Proto-NigerCongo reconstruction *za/dza “blood.tj=sn Hr Sps pn mr=Tn wAH n=Tn Wsr nb anx who will pass this tomb-chapel. J. Ahlo obidza. As we can see. Leipzig. Aegyptische Lesestücke. If we are equating Yeshua with Wsr/Jsr. the blood is soul. whom God hath raised from the dead. whom God raised from the dead.root in Wsir. New International Version (©1984) You killed the author of life. just like Wsr. whom God raised from among the dead. Douay-Rheims Bible But the author of life you killed. Sumerian sa. Lefana ubu-dza. Hinrichs'sche Buchhandlung. whereof we are witnesses. of which we are witnesses. the essence of Wsr is based off the -s.C. some renderings use the term “lord” while others use “prince” or “author. p. 1984: 43). as you wish that Osiris is merciful to you. This root can be seen in Yorùbá Ò jé “He was/He is – He happened to be.” We note that Eṣu among the Yorùbá is also known as a prince (Thompson.root which essentially means “to activate. whom God hath raised from the dead. In Yorùbá we have sí “to be” which is cognate with Hebrew yesh “there is” (=Assyrian 46 K. English Revised Version and killed the Prince of life. Kilimonjaro (Caga) samu. of which we are witnesses. every lector priest and every official swA. Below follows different interpretations of this passage: Bible in Basic English And put to death the Lord of life.It is clear that both Wsr and Yeshua are associated with living water. then the epithet “lord of life” can further be substantiated by examining the following Egyptian texts: Address to the visitor of the tomb of Neferenii j anxw tpjw-tA sS nb Xrj-Hbt nb sr nb O living who are on earth. but God raised him from the dead. I believe we can associate the Yorùbá terms Òjè “life” (after death) and èjè “blood” (êjë “vow. Blood As stated before. the lord of life The actual epithet “lord of life” is applied to Yeshua in the Bible in Acts 3:15.” Blood is life. We are witnesses of this. Since both are the embodiment of this element. Yorùbá e-dze “blood. Ewe ku-dze.” Nupe has e-dza.” blood oath?) as related to our -s. Guang obu-dza. of which fact we are witnesses.” This /s/ alternates with /j/. 88. every scribe.” Where there is no water. sau. 1924.” jí “Wake up from sleep” (back into consciousness and life). Darby Bible Translation but the originator of life ye slew. No matter how you translate it. whereof we are witnesses. Sethe. 67 . it was probably wine at the Last Supper (the Bible doesn’t mention the type of drink) that was to be symbolic of Yeshua’s blood (1 Corinthians 11:23-26).” The tit has been displayed alongside the anx Dand the DdD pillar signs as early as the third dynasty. Ti.su “to be.” In terms of Egyptian literature. I eat in it. The hieroglyph is usually translated as “life” or “welfare. the feminine side of Wsr. Kúda yi si “Death kept this one alive. This is comparable to Jesus in the Bible who is described in Revelations 1:15 as having “feet of fine brass. (Ashby 2001: 205) – Prt m Hrw of Initiate Anhai Blood is a very prominent symbol associated with Jesus in the Bible.” Budge interprets this sign as “an amulet symbolic of the uterus of Isis” (Budge 847b). I drink in it. to have”). Isis blood The ti. This word may have derived from a ts or dz root which would be in alignment with the earlier mentioned terms (*za/*dza) meaning “blood” (life).” The term sí means to have “life. It should be noted that Shu is associated with blood as well as can be seen in the following: I am one who is known by the god Htp.t = amulet. and from red glass. I have attained this field of yours.” it is often depicted in a red color and was often fashioned from red stones such as carnelian. O Htp. I will not be agitated in it for I know the wooden post that is in Htp.” in other words.” to “exist. Because it is associated with “the blood of Isis. It is called Bqwtt. jasper. The origin of this emblem appears to be unknown and may be a variation of the anxD symbol.”48 This -s. thet G. It will be demonstrated later that Wsr is also associated with the god Shu in ancient Egyptian mythology.47 We should also note that among the ancient Greeks the term zoe (s>z) is “life personified” (usually paired with logos “the word”).root is found in the Yorùbá phrase dá wa sí “Keep us alive” (Preserve us). Besides the symbolic association of blood and “cleansing” of sins (John 1:29). “to be.t G is known as the “knot of Ast. 1993: 143). One of her emblems is called the tit. However.t/Isis is. In Kiswahili eshe also means “life. I plough in it. It was made steadfast with the blood of Shu and held fast by the sweetness of the day when the years are separated. tiyet. Wsr is not directly associated with blood (to my knowledge).t = amulet Ti.” 47 68 .t’s “arms” are folded downward and the “arms” on the anx are raised horizontally. Brass It should be noted that the Yorùbá goddess Oṣun is represented by brass beads worn around the neck among her devotees (Neimark. I copulate in it. The difference between the two being that the ti. As. Fish Another comparative point of Oṣun is that her divine messengers are “fish” (<sin “a kind of fish” in Yorùbá). Jesus, the divine messenger, is also symbolized by modern Christians with a fish icon. . In Luke 24:42, after Jesus resurrected and met-up with the disciples, he asked for something to eat and “they gave him a piece of a broiled fish, and a honey-comb,” for which he took and ate in their presence. Some translations omit the honey-comb. None-the-less, Oṣun is also associated with honey (Neimark 1993, Epega 2003) and it is given to her as an ebo (sacrifice, offering). It should be noted that Wsr was often given honey as an offering and the Israelites in their early days gave honey as an offering to their gods.49 Peacock We also note that the peacock is symbolic of Oṣun among the Yorùbá. It is symbolic of her beauty and bearing, and peacock feathers adorn the thrones of Oṣun (Neimark, 1993: 143). This is important for our conversation here because ancient people believed that the flesh of a peafowl did not decay after death, and it became a symbol of immortality. This symbolism was adopted by early Christianity, and thus many early Christian paintings and mosaics show the peacock. The peacock is still used in the Easter season especially in the east.50 I don’t see this as having a Biblical origin, but it is interesting the parallel here and it is worth mentioning. Sacrifice Hieroglyphic renderings of Hapi(w)/Apis Wsr is also associated with the god Hapi (Apis), a black-bull that was sacrificed on behalf of the Egyptian community when it reached age 25 or 28. When an Apis bull died, the body was embalmed and entombed with the great ceremony that would be afforded royalty. This sacrifice was also a part of the Sd festival believed to rejuvenate the king’s vital power. Hapi is attested since the 1st dynasty according to the 5th dynasty Palermo stone, but not much is known to us before the New Kingdom period. His cult center was located in Mn Nfr. Greek and Roman authors have mentioned much about this Egyptian deity and inform us that it was conceived by a ray from heaven; other texts infer lightning. The Apis bull was originally the herald (wHm) of Ptah and was said to be born Some translations say “bronze” (like the New Living Translation and the New International Version Bibles). See Eva Crane The world history of beekeeping and honey hunting. Routledge, 1999 , p.595 50 "Birds, symbolic." Peter and Linda Murray, Oxford Dictionary of Christian Art (2004). 48 49 69 of a virgin cow51, impregnated by the spirit of Pth “break open, open to work, carve, god of craftsmen” (Tshiluba PaDika “open, burst, tear,”; PaTuk(a) “spring forth, arise, out”; KuPanda “build”; Pandish “save,” Mu-Pandish “savior”; Amarigna baTS’hae “carve, engrave”; Kiswahili patua, pasua “to split open”). After the death of Hapi (Apis) he fused with Wsr and became known as HpWsr. Later during Hellenistic times, he became associated with the Greek god Serapis. The Greek amalgamated god Serapis which combined components of the Greek gods Zeus, Asclepius52, and Dionysus as well as the Egyptian deity Osiris and the sacred Apis bull cult. What’s important for our purposes here is that the bull Hapi, who is associated with Wsr, was sacrificed for the people of Egypt just like how Jesus was sacrificed for the Israelites. Not only was Hapi sacrificed, but like Jesus of the Biblical tradition, was born of a Virgin mother. Like Wsr, Apis represented the life-force in nature and procreative power. He was called the Ba (soul) of Wsr. Another striking correlation with Jesus the Christ is that the Apis bull was often mummified53 and fixed in a standing position on a foundation made of wooden planks. Remember that the Greek word in the Bible rendered in modern times as “cross” is stauros which is a wooden pole, or stake. Certain scriptures, however, state that Jesus was hung from a tree. Thomas O. Lambdin (1983: 307), Introduction to Sahidic Coptic, notes that a word in the Sahidic dialect of Coptic for “cross,” as in wooden stave, is she, shay or shi which sounds much like our -s- root we have been speaking about throughout this text. This may be relevant here. Wilkinson (2003: 170) Asklepios himself was patterned after the living philosopher, doctor and architect of Egypt (3 rd Dynasty) named Imhotep, who himself was later deified among the Egyptians. The above image of Serapis courtesy of. This image looks strikingly like the modern depictions of Jesus the Christ. 53 To see one of the mummified bulls online, visit the British Museum’s website at:. It should be noted that Jesus was also “mummified” in a similar manner as he too was wrapped in white linen and a shroud (Matt. 27:59). 51 52 70 BREAD OF LIFE In John 6:35 the Bible informs us that Jesus is the “bread of life.” This is interesting for our purposes as Wsr is depicted as the soil from which wheat (the material needed to make bread) is grown out of. He is also symbolic of the wheat itself as can be seen in the image below. This relief above also demonstrates another association of Wsr. As we know wheat grows out of the ground and in the picture above it is growing out of Wsr’s body. We can safely assume here that Wsr represents the nutrient rich and fertile soil (which is why he is often depicted as pitch black). This would explain why he is deemed “lord of the underworld.” The “underworld” is simply a modern term for soil, the ground. Yorùbá siri "a stalk of corn or rice with grain on it," is relevant here as Wsr is also the god of corn.54 Bread is made of wheat and we should all be familiar with the Bible’s association of bread being the body of Yeshua. In 1 Corinthians 11:23-26 (King James Version) it states the following: act of eating the body of Christ is known as the Eucharist ritual in Biblical literature. One would argue that Wsr representing wheat and Yeshua representing bread is pure coincidence. However, Wsr also represents bread and his devotees also ate the body of Wsr as a religious ceremony. See chapter 40 of James George Frazier (1922) The Golden Bough: The Magic Art and the Evolution of Kings, 3rd Edition. 54 71 The Ancient Egyptians celebrated a “mystery” festival in honor of the death and resurrection of Wsr who was murdered by his brother St (Sutekh). The day this was celebrated was also the same day that the grain was planted in the ground by the farmers. A play was conducted in the temples by actors who would re-enact the murder of Wsr by St.). There were two aspects to the plays for this ceremony: a public one and a private one which only the priests in the temples participated in and initiates witnessed. The details of these rites are discussed in the I-Kher-Nefert stele erected in Abjw (Abydos)—the primary city where Wsr was honored—during the 12th dynasty (1875 BCE). Plutarch (born 46 CE) mentions that clothe and adorn, this indicating that they regard these gods as the substance of Earth and Water." (Isis and Osiris, 39) In the Osirian temple at Denderah, an inscription (translated by Budge, Chapter XV, Osiris and the Egyptian Resurrection) describes in detail the making of wheat paste figures/models of each dismembered piece of Wsr to be sent out to the town where each piece was discovered by Ast.55 At a temple in Mendes, figures of Wsr were made from wheat and paste. These were put on a trough on the day of the murder and water was then added for several days. This carried on until finally the mixture was kneaded into a mold of Wsr and taken to the temple to be buried (the sacred grains for these cakes were grown only in the temple fields). Each mold for the 16 (or 14 pieces) of Wsr’s dismembered body were shaped from a red wood tree (save the phallus). The cakes of bread baked from these molds were then placed in a silver chest and set near the head of the god with the inward parts of Osiris as described in the Book of the Dead (XVII). These rituals were climaxed by the actual eating of the bread, Wsr’s body, where the devotees were symbolically transformed into Wsr. This festival survives among the Yorùbá as reflected in the term ṣorò "to observe a festival of a god" which maintains our s-r root mentioned throughout this discussion. One festival that comes the closest, lexically, in Egypt is the sn.t festival which is a celebration of the 6th day of the lunar month (s-n <s-r?). Although the way the rituals are carried out in Egypt and in the Bible were different, the correlation is with Wsr and Yeshua both 1) being associated with “bread” and 2) both were symbolic of their bodies. Not only that, it was required in both traditions by the priests (disciples) to eat the respective bodies. These aren’t “coincidental” correlations and no amount of downplaying the exact carrying out of the story is going to diminish the facts presented concerning this ritual. The Eucharist rituals predate the Bible, the Hebrews, let alone Yeshua the Christ. Jesus was, in one aspect, the personification of ancient agricultural phenomena and the associations are readily apparent in the New Testament as we have seen thus far. In the myth of Wsr, his brother St/Set sliced his body into 14 pieces and scattered them across Egypt. It was Wsr’s wife Ast that located the pieces and remembered them. 55 72 THE DIVINE SHEPHERD Yeshua is often depicted as a shepherd. This is something we would expect from a culture that has a long history of being pastoralists. However, Wsr is also a titled the “good shepherd” and it is evident in the royal emblems he carries, especially the shepherd’s crook. Wsr holding shepherds crook and flail. Osiris was also called Asar-Saa (Murdock, 2009: 313). The word sA means “protection, magical protection, talisman, amulet.” Other related terms are: Tsw bsA Sw stp sA Sn sA sAw sA/sA pr hsw “commander, protector” “protection” “protection, shade, screen” “protector” “protection, ring of protection” (a circular sign) “regiment of troops” (the army protects the citizens) “guard, restrain, ward off, heed” “pasture ground, byre, cow shed” “magic spell” (for protection against crocodiles) A shepherd is a protector of cattle or livestock. Given our criteria concerning the relationship between the occupations and characteristics of the main characters of myths, we would expect that somehow the s/s-r root will be found in association with a shepherd. The Ancient Egyptians never let us down. The word for shepherd in Egyptian is mniw siw.t. It is composed of two words: mniw "herdsmen, shepherd," mniw "herdsmen" (which itself derives from mni "to guard") and siw.t "sheep." A shepherd is simply “one who guards sheep/the herd.” In other words, a shepherd is a protector which is synonymous with a “savior” (a shuwa’). A variant of the term for shepherd in Egyptian is mniw anx "goatherd." The word shepherd can also be rendered awti "shepherd, crook bearer." In keeping with the s-r root, the word for “ram, sheep” is Sr/sr. Jesus is known as the “lamb of God” and a lamb (a baby sheep) would fall into our Sr/sr root. As we can see, it was practically inevitable that Jesus would be considered a shepherd (and the lamb); not because he existed and actually was one, but because he is a mythological figure who has taken on the attributes of terms based on the s-r root embodied in Wsr “the shepherd.” 73 The son of Wsr, Hrw, was also known as a “good shepherd” in the third inscription at the Temple of Redesiyeh or El Tadesia at Wady Abad, near Edfu in Upper Egypt.56 As John P. Lundy stated, “The royal Good Shepherd is the antitype of Horus” (cited in Murdock, 2009: 312). The idea of the HorusKing as the “good-shepherd,” in fact, was so important that it constituted a major shift in perception and public policy, representing the general mentality of the 11th and 12th dynasties (c. 2050-1800 BCE). Egyptologist John A. Wilson, who is the director of the Oriental Institute at the University of Chicago, informs us that, “The concept of the good shepherd rather than the distant and lordly owner of the flocks shifted the idea of kingship from possession as a right to responsibility as a duty” (ibid.). In other words, the association between the king and the shepherd had already taken place in Egypt 2000 years before Jesus was alleged to have lived. If the Israelites were in Egypt as their myth alleges, then they definitely would have adopted this conceptual analogy in regards to kingship. Remember, the Israelites did not come from a tradition of kings. They adopted this form of rule based on what they witnessed in places like Egypt. One of their leaders warned against it, but eventually the Israelites got the o.k. to be ruled by a king. It should be noted that until that point, there was never any association of God having a “throne” or being considered a “king.” Eṣu the Good Shepherd It is important to mention here—as pertaining to our correspondences between the Yorùbá deity Eṣu, Egyptian Wsr and Hebrew Yeshua—that the Yorùbá deity Eṣu also carries a shepherd’s crook in the tradition of Ifa. The crook is called a pankere57 and it is used by Eṣu (and elders in general) to correct a child for a mistake he or she has made or to beat an animal. Eṣu is a deity that punishes bad behavior due to bad character. Jesus ultimately plays this same role as he is the “god” of judgment who will come to punish in the times of Revelations. This is something that cannot be down played: by either Egyptologists or Biblical scholars. This is evidence of a long standing tradition that has nothing to do with Christianity. Yorùbá king holding crook in left hand and flail in right hand. 56 57 James Henry Breasted. Ancient Records of Egypt, III (2001), University of Illinois Press. p. 86 Baba Sangodare Fagbemi, priest of Ifa; personal communication (January 2011). Cairo, Egypt 74 Yorùbá kings also carry the shepherd’s crook and flail royal objects like the ancient Egyptian kings as can be seen above. Eṣu was once a prince and the crook has been a part of the royal regalia ever since, in honor of their nomadic, pastoral roots. 75 THE PASSION Many are familiar with the brutal murder of Yeshua in the Biblical myth. Others may be familiar with the murder of Wsr by his brother Set in the Egyptian drama (see Muata Ashby 1996, 2001). In this brutal murder, Wsr was cut into 14 (some say 16) pieces and his body was scattered all across Egypt. Everywhere that a body part was found, a temple was erected in his honor. But many may be unaware that Eṣu among the Yorùbá tradition also suffered a similar fate as Wsr. Robert Farris Thompson relays a story concerning Eṣu that has a similar undertone as the story of Wsr in ancient Egypt. Instead of temples being erected in the case of Eṣu, his body parts became laterite stone shards which diviners use as representative of his body. The cone of laterite (yangi) appears in Yorùbá markets over which cult officials pour daily offerings of palm oil to maintain Eshu's problematic coolness. Laterite-cone altars to Eshu recall a myth whereby Eshu devoured enormous quantities of fish and fowl offered to him by his mother, and finally devoured his mother, too. Where upon his father, the god of divination, alarmed, himself consulted a divination expert and was told to sacrifice a sword, a male goat, and fourteen thousand cowries. The god of divination did as he was told. Consequently, when Eshu threatened to devour him, too, the god took a sword and hacked him to pieces, and the pieces became individual yangi, lateritic shards. Orunmila pursued Eshu through nine heavens until finally, in the last heaven, Eshu was pacified by Orunmila and said that all the particles of his spirit, the yangi stones and shards, would become his representatives. All Orunmila had to do was to consult them (make sacrifices upon them and ask a blessing) whenever he wanted to send them on a mystic mission. Eshu then returned his mother, alive, to the world. (Thompson, 1984: 21-22) This may be further evidence of an ancient shared tradition. It is interesting, that in three different parts of the world, three deities, with the same name and attributes were all “cut-up” and killed in each one of their mythologies. Eṣu and Wsr are the only ones out of the three who had a full body hacking, while Yeshua was “sliced” and lashed with a whip (similar to a cat-o-nine-tails). Further investigations will yield better information for which we can better interpret what these “cuttings of the gods” mean. Space will not allow further exploration here. 76 I AM THE WAY – THE CROSS AND THE CR OSSROADS One of the most significant symbols of Christianity is the cross which has become a visual metaphor for Jesus’ suffering and alleged lynching on the cross. There has been much debate on whether Jesus actually died on the cross as a few verses in the Bible state he was hung from a tree. Acts 5:30 Acts 10:39 Acts 13:29 1Peter 2:24 Galatians 3:13 “Jesus, whom ye slew and hanged on a tree” “whom they slew and hanged on a tree” “they took him down from the tree” “who his own self bare our sins in his own body on the tree” “Christ…being made a curse upon us…Cursed is every one that hangeth on a tree.” The word for “cross” in Greek is stauros #4716 or stauroo #4717 in the Strong’s Greek Dictionary. But this has proven to be a later rendering of the term, not the term used in the Bible. Vine’s Expository Dictionary of New Testament Words states stauros denotes, primarily, an upright pole or stake. Regardless of the controversy, the cross is a symbol of Christianity now. Thomas O. Lambdin (1983: 307) Introduction to Sahidic Coptic notes that a word in the Sahidic dialect of Coptic for “cross,” as in wooden stave, is she, shay or shi. This may be relevant in regards to our s/s-r root. The cross symbol is actually pre-Christian and is used as a spiritual sign in many cultures. For example, the god Bacchus is shown with crosses on his headdress. In 134 BC in Syria, it was found that the cross was recognized as a symbol of victory over death (Biederman, 1992). A white-robed heavenly Egyptian “Brotherhood” with Cross Hieroglyph (Hornung, The Vally of the Kings, 160) In Africa the symbol of the cross is a spiritual sign which denotes the crossroads, the demarcation point between the material world and the spiritual world. It is the location where sacrifices or offerings are made. When drawn on the ground, it is used as a location to swear one’s innocence on or to make oaths (Thompson 1984, Halloway 2005: 288-289). All of these rituals are associated with the Christian cross as Jesus was the literal sacrifice, where his blood was used to pay a debt on humanity’s behalf and to this date offerings are made in his name to the cross. It is our contention here that this is just an adaptation of old African practices where instead of an animal being sacrificed on the cross (symbol of the crossroads) it is Jesus (the lamb of God). What is worthy of note here, as to be expected by now, is that Wsr of Egypt and Eṣu of Yorubaland are both entities associated with the cross (crossroads) before any interaction with Christian 77 Europe. In the Old Kingdom Egypt, the deity Anpu (the opener of the way) used to be over the funerary offerings of the deceased. But by the 5th dynasty Wsr had taken a more prominent role which is evident in the standard opening in the funerary texts which stated, “An offering which the king gives to Osiris [on behalf of the deceased]…” (R. Wilkinson, 2003: 122). In other words, Wsr is the spirit that presides over the offerings for the community (like the priests of Levi). It is through Wsr that the blessings are channeled and prayers sent to their appropriate place. Painting from the Tomb of Rameses I, Valley of the Kings, West Thebes, in Egypt. The cross “X” symbolizes that the person is deceased and they are in the underworld (the inbetween of heaven and earth). This suggests that before Yeshua the Christ among the Hebrews (Matthews 11:28-30), it was Wsr whom one was to bring one’s burdens to. In the pyramid text of king Unas of the 5th dynasty, Unas brought the evil sayings and strifes that his enemies directed towards him and it was given to Wsr to take care of. The text states: O Osiris all that is hateful in Unas hath been brought unto thee, and all the evil words which have been spoken in his name. Come, O Thoth, and take them unto Osiris, bring all the evil words which have been spoken and place them in the hollow hand; thou shalt not escape therefrom, thou shalt not escape therefrom. (Budge, 1967: CXXXIX) A more up to date rendering of this text reads: Utterance 23 16: Osiris, seize everyone who hates Unas, -libationwho speaks evil against his name! Thoth, go, seize him for Osiris! Bring the one who speaks evil against the name of Unas, place him in your hand! To say four times: "Do not separate yourself from him, 78 take care that you are not separated from him". 58 The overall spirit of the excerpt is that Wsr/Osiris is the one who fights on behalf of the devotee. The “X” symbol on Wsr’s chest, coupled with the fact that he is lord of the dwAt, means that these burdens are presented to him at the “cross” which is symbolic of the realm for which doors (opportunities) are opened and closed. In keeping with our cross-African comparisons, Thompson discusses Eṣu figurines that have a sharp, blade-like projection, or knife, on the top of Eṣu’s head. He tells us that this aspect of the carving (the pointed head) symbolizes that Eṣu cannot shoulder ordinary burdens (Thompson, 1984: 28). With Eṣu being the crossroads, it is clear that the most extraordinary burdens of the people are taken to him to be dealt with, just like Yeshua of the New Testament and Wsr of ancient Egypt. We will see later on how this concept relates to an Nganga among the Bantu and the nkondi figures (which symbolize the crossroads) which represent the initiated priests who take on the community’s burdens. In the same text Thompson discusses a song that was sung in places like Havana, Cuba, New York City and Miami by Eṣu devotees whose lyrics concealed a visual pun on the single crimson feather that Eṣu wore erect upon his head in the presence of the Almighty. Both feather and knife are described as preventing Eṣu’s head from being used to support an ordinary burden (ibid. 28-29). He noted that it was inevitable that nails and other objects would be used to simulate the knife atop the head of Eṣu in the clay and concrete sculptures of Havana and Harlem. Although these were practices recorded in the New World, the knife on top of Eṣu’s head is an original Yorùbá practice. I mention this here because it reminds us of the image of Jesus’ crucifixion when he had to wear a crown of “thorns.” If our overall correlations are sound, which we think they are, then we can posit that the thorns on Yeshua’s head were symbolic of the extraordinary burden he had to bear: the sins of humanity which was “no oridinary burden.” Eṣu is also known as eṣuona “Eṣu of the way,” the opener of the path to the mysterious source of knowledge (Ford, 1999: 157). It is for this reason that Eṣu is intimately connected with divination, as he is the channel for which the divine message is sent. This makes Eṣu, like Jesus the Christ, the divine messenger. The crossroads symbol is also utilized in many diagrams used for divination. Whenever traditional Yorùbá encounter change or challenge in the world of Eṣu, the limitations of individual calm and wisdom become acute (Thompson, 1984: 33). For such cases, a person then relies on the accumulated insights of the poetic chants of the Yorùbá divination system called Ifà (the science of change) to place his or her individual problem in perspective. A Babalawo (father of the mysteries, a diviner) uses an instrument called an Opon Ifà which the crossroads is an important aspect. 58 From the Sarcarphogus Chamber, North Wall, First Register: (Retrieved November 30, 2010). See also James P. Allen The Ancient Egyptian Pyramid Texts. Society of Biblical Literature. Atlanta, GA (2005: 19). 79 A Yorùbá Odu chart based on Ifa cosmological wheel which includes the four elements: air, water, earth and fire.59 The crossroads inside the Opon Ifà symbolizes the cosmic life cycle, the four cardinal points (north, south, east and west) and the four elements (air, water, fire and earth) (Epega 2003, Neimark 1995). In Yorùbá cosmology, the Opon Ifa symbolizes the universe which is denoted by the Heaven (orun) and the Earth (aye). The heaven represents the masculine force which directs the beginning of things and the earth represents the feminine force which gives them their completion. The crossroads is the center point where heaven and earth meets and it takes a specially trained priest, who has access to both worlds, to facilitate this cosmic dialogue. The cross is not only a symbol of offerings, divination and death, but is a sign used to denote a priest and healer called an nganga in central, east and southern African societies. In order to better understand this symbolic association, a brief explanation of the Bantu cosmology is in order. It is through the Bantu cosmology that the conceptualizations, epithets and motifs associated with Wsr of ancient Egypt and the Hebrew Yeshua begins to make sense. Ankh – Nganga – And the Cross The Bantu-Kongo cosmogram (dikenga) expresses an ancient teaching that undergirds the collective African world-view. Fu-Kiau (2001: 127-50) dedicates a considerable amount of space to discuss this ancient Bantu secret that was introduced to him in the Lemba Institute60 in Manianga Lower Congo where he was born. In cross comparing African cosmologies, we come to discover that the model for this secret teaching may help to explain many phenomena we witness in the ancient Egyptian texts. In African societies there is no ultimate death inside the infinite universe. Energy can only change form and perpetual evolution is the law within the infinite. The living sun (the human being) is on an eternal spiraling course that is forever rising and setting in the upper and lower worlds respectively (see Dikenga diagram below). The cycle of life articulated in the cosmogram shown 59 See Baba Medahochi Kofi Omowale Zannu Mogbarimu Ajinaku El‟s The 16 Great Signs and their Meanings. The Akoda Institute. Atlanta, GA. 60 This was one of five institutes of ancient Kongo. The other four are Kimpasi, Kinkimba, Bwelo and Kikumbi. 80 below is inspired by the “apparent” movement of the sun around the earth. It would appear that all major African spiritual systems are rooted in this concept. The circle is divided into four primary demarcations (n’kama), each representing a major stage in the development of material and world systems: the nseluka (“sunrise”), kunda (“zenith”), ndima (“sunset”) and n’dingu-a nsi (“midnight” or “depth of the world”). These stages are represented by geometric “Vees” that represent living pyramids in motion (Fu-Kiau 2001: 12). These stages of development are not only associated with human life, but with all political, spiritual, communal and cosmic life cycles in the universe. Anything that has a finite existence follows this path [dingodingo]. These four stages are represented by four esoteric colors: Musoni – Yellow Sun – Sun of perfection Kala – Black Sun – Sun of vitality Tukula – Red Sun – Sun of maturity, power, warning, and danger Lupemba – White/Grey Sun – Sun of death and change These phases and colors are also used to represent the formative stages of the universe and “earths” (planets). The musoni sun is so-called because the color of this position is musoni (yellow ochre).61 This represents the deepest mystery of the universe and its origin. The planets formed at this stage in universal evolution. They were musoni worlds because of their burning/radiation. After the burning process cooled down totally, it gave way to the kala sun (literally, “the sun under which one can be or exist”). Kala, meaning charcoal or black, is the symbolic color of this era.62 Life becomes possible at this stage of development in the physical world because of the green life present (gheghe, mindangwa).63 The growth of life began in this era. The kala burning sun era was followed by the tukula sun era (the red sun). This is the highest level that can be reached by any living entity (civilization, technology, etc.). This is the stage of maturity and power, but is also a warning and danger era. With this being the height in the evolution of life, what follows is death and decline of the life form. Animals, according to Bantu cosmology, came into being during this era. The tukula era was followed by the luvemba (mpemba) sun era (the era of greyness). This is considered a negative element and causes natural physical death. Luvemba represents the era of the greatest danger to all growing life. It is the declining stage of development. Ironically, according to Bantu teachings, the human being developed in this stage. The “V” is the basis of all realities; but to understand this cosmological concept, it was necessary to explain first the Bantu cosmology to provide context for the discourse below. The “V,” in part, is named after the “V” shape that each section of the cosmogram forms. 61 62 Fu-Kiau (2003:65) For a full break-down of this term and its use in the proposed name for African-Americans, see The Bakala of North America: In Search for a Meaningful Name for African-Americans by Asar Imhotep (2009) 63 See Fu-Kiau African Cosmology of the Bantu-Kongo (2001:23). 81 The most important stage that concerns us here is V-3 (Vanga) which represents the Tukula (red colored) sun of maturity, leadership and creativity. The word Vanga derives from an archaic Bantu verb ghanga which means to do, to perform. It is where we get the term NGANGA which means a master, a knower, a doer, a specialist, etc. Most text books render this word to mean “doctor,” but that is just a surface understanding of this label. In Bantu languages one can often turn a verb into a noun by means of affixation. In this case the letter -n- is a contraction of the word ENIE or ENYIE which means “one who, a possessor, that which.” In the case of the Kikongo term N-GANGA it is a statement simply saying “one who does” or “one who performs.” This Vee is the most critical in life as it represents the stage of creativity and great deeds or tukula stage of the root verb kula which means to mature or master. What is implied by this term is that an nganga is someone who is highly knowledgeable, highly respected and also a community leader who has put in work to enhance the lives of the community and to maintain balance (to make sure the community waves aren’t shaken) of village life. An nganga is a master, a doer and a specialist in a community of doers. Dr. Fu-Kiau expounds on this subject and stage of development more fully in his African Cosmology of the Bantu Kongo. He informs us that: This Vee, the third, is a reversed pyramid. It occupies the position of verticality [kitombayulu], the direction of gods, power and leadership. People, institutions, societies and nations as well, enter and exist in this zone successfully, only if they stand on their own feet. One enters and stands up inside this Vee to become a doer/master [nganga], to oneself first before becoming an nganga to the community (…) To stand “well” inside this scaling Vee is to be able not only to master our lives, but to better know ourselves and our relationship positions with the rest of the universe as a whole. (Fu-Kiau, 2001:140-1)(emphasis mine) An nganga is an initiated master. The word for initiation and the word nganga are similar in morphology. The word ghanda means initiation. It is similar to the word ghanga which means to perform or to do which becomes nganga: a master, a doer, specialist, community leader. This 82 directly informs our discussion in many ways. The first is linguistically which we will discuss further below. The second is iconographically as Dr. Fu-Kiau plainly informs us that: This power figure, the leader/priest [nganga], who stands powerfully at the center of the community issues [mambu], became the Egyptian ankh or symbol for life. Of course, among Bantu people, an nganga stands “vertically,” and powerfully inside the community “Vee” [telama lwimbanganga mu kanda], as the symbol of active life in the community. (Fu-Kiau, 2001:131-2) (emphasis mine) As we can see in the account from Fu-Kiau, the anx symbol, in the Kongo context, means more than simply life: it represents an active life of a master; thus ghanga (to do, to perform). It is my contention that the actual pronunciation of the word anx is actually closer to the pronunciation of the word for life and man in the Bantu and the Akan languages: nkwa. At the heart of the word nkwa or anx is the word ka or kaa which in the Egyptian and Niger-Congo languages mean fuel, power, life, to have life, be, have being, spirit, energy, be burning and more. The Bantu cosmology, symbolized in the Dikenga cosmogram above is going to be key in interpreting some other concepts below. It is this cosmogram, and its corresponding symbolisms, that help us to accurately interpret certain poses in Egyptian reliefs. The following is an excerpt from an online article I wrote titled Posture and Meaning: Interpreting Egyptian Art Through a Kongo Cultural Lens.64 In this article we analyze certain poses and their meanings in the culture of the Kongo and compare them to Egyptian motifs in hopes to better understand the meaning behind Egyptian postures. A key pose of the Kongo that is important to the topic at hand, is the Crossroads Pose: 64 83 The Crossroads Pose The crossroads pose, with right hand up to heaven and left hand parallel to the horizon line, characterizes the niombo figure. The niombo of the Kongo would be equivalent to the orisha of the Yorùbá, or the neters of the Egyptians. This gesture is found in Kongo, Cameroon and Nigeria and has made its way to Haiti via vodun. The right hand up and left hand down recalls the anthropomorphic reading of the hand-guards of the mbele a lulendo (knife of authority), the royal swords of execution in ancient Kongo. In Kongo this gesture, on the swords and niombo, marked the boundaries between two worlds (upper and lower worlds). We see this pose on another version of the nkondi figure. We will also observe that the ancient Egyptian Bes is also in this position. Nkondi65 (Kongo) with crossroads pose Egyptian Bes with the crossroads pose 65 Unknown Kongo artist and ritual expert, Democratic Republic of the Congo or Angola, Nkisi nkondi power figure, about 1890, wood and mixed media. Purchased through the Mrs. Harvey P. Hood W‟18 Fund, the William B. Jaffe and Evelyn A. Jaffe Hall Fund, the William B. Jaffe Memorial Fund, the 84 This pose relates to the cosmogram mentioned earlier and could signify in the Egyptian the meeting place between the ancestral realm and the manifest realm. Remember that an nganga. Now examine the following predynastic Egyptian image with the same pose: We briefly discussed earlier the relationship between Wsr, Eṣu and Yeshua’ and the concept of “life.” The African cross, called the anx D in Egypt, is the symbol for life and whose word actually means life (Tshiluba anga “life”; Kikongo, Akan nkwa “life”). In Christianity, the cross is a symbol of Jesus’ victory over death. It is my contention that this is a simplistic view and one that is expected as most Christians are not aware of the “concealed rear-view” message behind the symbol for which only an African-Centered perspective (and African languages) can decode. The root anx in Egyptian literature is usually given the following meanings: ANKH → life ANKH → live, life, be alive, oath ANKHW (NKWA?) → the living William S. Rubin Fund, the Julia L. Whittier Fund and through gifts by exchange; 996.22.30233. On view in the exhibition Art That Lives? 85 ANKH → person, inhabitant, citizen, living one ANKH → person, citizen, living one As Fu-Kiau noted, the anx symbol does not only represent life, but the life of a healer, a master, a doer in a community of doers. In other words, this term and symbol is associated with the wise men and women of the pre-Western communities of memory. It is a symbol, borrowed from the anatomy of man, which represents the ideals of African centers of wisdom: an emblem of the priesthood and the priests. It is no coincidence then—given the correlations we’ve seen thus far and those to follow—that this sign is also associated with Jesus the Christ: a Jewish priest. We note its association with Wsr who is often depicted in the anx pose. In the Kongo, for instance, it is a symbol of all men and women who have demonstrated a high level of wisdom, maturity and service to the community. It is the highest point of achievement in African communities of memory. NKWA (BaKwa “people”) Ancient Egyptian ankhu = man Budge 124B The name for cross in Hebrew is tslav and is not cognate with our term anx. This is to be expected as this is a symbol that was borrowed and adapted into Hebrew culture: it is not native to Hebrew people. They in turn used a native word to describe an African phenomenon, just as the Greeks did with the Egyptian tekenu which the Greeks called an obelisk (a phallic symbol representing rejuvenation and fertility). With this said, all of the concepts associated with the cross in the Christian tradition is represented by various terms, whereas among African people who speak the Niger-Congo languages, all of the concepts we associate with the cross, and the one who died on the cross, are all captured using the one root term: anx. Because mainstream Egyptologists have not done cross comparative studies with Sub-Saharan African cultures, they have over simplified the meaning of the word anx and relegated it primarily to mean “life.” The term has a richer set of meanings and the field of Egyptology would benefit from turning into inner Africa to gain clarity of the meanings of the motifs present in ancient Egyptian iconography. The word anx is used in many words in the Tshiluba language and all sheds much light into the actual symbol of the anx which is often depicted as a knotted rope. We begin our analysis by observing a few terms that relate to the nganga (whom Jesus is a personification of) and the cross to get a better understanding of who Jesus was and why the cross had to be associated with his divinity. 86 Terms with the root anx in the Tshiluba-Bantu language66 -Sanga di-/ma-Sanga ma-Sang(u,o) mu-/mi-Sangu muSangu sangisha Sangila, diSangila, Cisangilu sangala sangala -sàngalala -sàngaja -sàngalaja Sangasanga Sangesha Mu-Sangelo to cross, join together, gather, mix crossroads, confluence, junction, joint, node interval between the shoulders; junction point of two shoulders or two arms = nTangani/diTung (U, A) - a-mapa/makaya/mapwapwa time, period congenital defect, pathology of birth to make gather, join together, mix Gathering, reunion, meet, communion, community to be delighted, to be happy, to be in good health to recover (health, vitality/strength joy, the calm one) to be full with joy, good fortune, to be completely happy to cure to cure, return health Joy To welcome good fortune, success, happiness, joy As we can see above, this root is associated with terms that relate to the “crossroads, to cross,” “good fortune, success,” “recovery of health, healing” “happiness, joy,” and the “uniting and gathering of a people.” In African communities of memory, it is the elders who call community meetings to discuss important issues that concern the community. These meetings are held at centers with various names: mbongi, boko, kioto, yiemba, etc. (Fu-Kiau, 2007). An nganga, a musangila would be one who calls for these gatherings.67 All of these concepts are associated Jesus the Christ. It is my contention, again, that these associations are not connected with Jesus because he was an actual person, but because he is associated with terms that unite these many layers that are in the associative field of meaning for our anx root (which also matches many terms with our s-r root). This is typical of myths and all of this can be seen with the associations of Eṣu and Wsr. The above s-g root (with nasalized g) is the same root (s-k) in Hebrew that gives us our term Christ in English: ma-shiakh (s-kh) “anointed.” This will be addressed in the next chapter. Here are a few epithets of Wsr that might need some reinterpretation. See Mubabinge Bilolo and Nsapo Kalamba Renaissance of the Negro-African Theology. Essays in Honor of Professor Bimwenyi-Kweshi [Renaissance de la Theologie Negro-Africaine Melanges en l’honneur du Prof. Bimwenyi-Kweshi] (2009: 131-135). 67 Yeshua of the Bible is known to gather hundreds, sometimes thousands at a time to hear his messages. 66 87 anx.ty "the living one" a title of Osiris (Budge 126a) ba anx "living soul" a title of Osiris of Tet (Budge198b) ba anx "a soul that has renewed its existence in heaven" babaw "soul of souls" a title of Osiris (Budge 199b) Ba-n-hh "everlasting soul" a title of Osiris Wsr with the anx “cross” pose Here are some alternate variations of the root: Shinga/Singa/Jinga MuShinga/ka-Nzinga muJinga muJinga Nshinga Zinga/Shinga MuShinga/Ka-Zinga MuShinga To bind, thread, roll up, knot, encircle, tie with a wire; to braid/weave Newborn with the umbilical cord rolled up Cord, reel; braided rope Invocation ritual Neck To wish, desire, prefer Child’s desire, wishes, expected Price, value, dignity = bunema Sanga is a synonym or variant of: Sanka Disanka Sankila Sankisha tunka Be happy, rejoice, have joy, to be delighted Happiness, joy, pleasure To be delighted for To please, make happy, rejoice Be happy, happy, joyful, dance of joy, rejoice, merry, exalter But S-NG- < Ś-H- is also the opposite of Sanga or Sangala Sunga diSunga Sunga Sunga kansunga-nsunga Sungula -sungila sungidila To hang; to strangle by tightening the throat or the neck (nshingu) To hang itself To separate (of the combatants), appeaser, to reconcile, pacify To choose, elect Partiality, bias To choose, select; to prefer; elected To deliver, protect from; to save To separate, deliver, defend against, to help The latter terms sungila and sungidila are important for this discourse because it is the root of the word Musungidila “savior” which belongs to the s-k root in Hebrew for the word ma-shiakh 88 “anointed”: a title attached to the Hebrew savior Jesus. For now, our anx root can be seen in other related terms that correspond to many of the characteristics of Jesus the Christ. Nanga >Dinanga >Dinanga Nanga >Dinanga Nanga NangaNanga Nanga-nanga Nanga Nangila > A-Nangila Nanga >Munanga Nenga Nenga di-Nenga Nengesha Nung > diNunga >diNunga >Mununga Build, fix, braid lock, knot into gear, resort, relaxation love, desire, prefer, desire = swa Love, charity = diswa overflow, be too hard = nangapala, nangisha Preferably, mostly love each other Beloved, darling = mu-/ka-/ci-Nanga Loving, to please, prefer, like most Who is most preferred, which surpasses all; that is too strong educate, exhort Exhortation constant thread = Nyanga Durer, endure, persist, continue, expand, persevere, persist The duration Sustaining Tie, plaiting, tying with a wire Node, wrist = dimungu Network, join Protector, guardian or messenger of the head As we can see here, although the prefixes are different in comparison to the previous tables, the meanings are essentially the same. This root is associated with a “teacher, a healer, something perpetual (eternal life?), a protector, a messenger, loving thy neighbor, charity, blessings and expansion.” All of these terms, based on one root, are associated with Jesus in one way or another. I argue that this is the case because the mythological Jesus was patterned after the priests from the living traditions of the Kongo via ancient Egypt. What was an everyday occurrence in Africa became a unique phenomenon in Palestine. I will digress here to continue our discussion. Another form of the root lends further insight: Nxa Nxa Nxa Nxa Nxa Nxa Nxa Nxa Nxa Nxa Nxa Nxa / n anx 68 Nke Nko Nko Nkonko Nkonka Nku niNga Nyunga Nyuka nyInka Nyeke Nyeke-nyeke Strong, tough, strong Fullness, achievement, perfection Abundant, many Sternum, ring/coil/link, collar Requests, questions, envy Violence of desire, fire, vehemence Twist, squeeze, bind strongly Move in a circle Toll-relaxes Ancestor, ancestor/grandfather68 = Kaku Nkambwa, Nkoko Perpetual, eternal perpetually Nyinka "ancestor" seems to have also the meaning of "eternal", "Imperishable Stars" 89 is a place for initiation: it is the Holy Land. for it is the source of life. The very word Kongo. "as a remedy against the dangers. primordial hill. It is my conviction that because Egypt was located in the desert. 90 . they brought the forest to them: symbolically. It is through initiations in the forest (the wilderness) that one discovers “God” and communes with the ancestors (who are believed in many traditions to live in trees). forests and mountain with sun over the horizon.Anga. We will see how this all comes together in the next chapter. The spirit world among those who live in the forest belts of Africa is seen as a forest. Kongo. The words cited below are the ones I feel speak to the purpose of this work in association of a Christ or messiah. remedy = the protected and cared for. Representation of Ipet-Isut temple breaking down the natural elements being represented in the temple construction: water. KoOnga. It is a place where initiates go to study firsthand the herbal plants needed to make medicines to heal those sick in their communities. It makes one to wonder as to why the temples in ancient Egypt were built with large pillars that looked like plants and trees. This is a testament to their ultimate origins and how environment dictated how the culture was to be adapted. Therefore.The meaning of life can spring from the Baluba's use of terms that mean "that which protects life". for which Central Africa is named. Present in: Ba-Kw. they could not go to the forests for initiations. Many of the terms cited above can be found in the ancient Egyptian language. KaAnga Remember our discussion of maanga in our section on African wisdom traditions." Anga > BuAnga/Manga Medicine. It is also a word for God. to perpetuate" Nanakana sangala sanx. 69 Retrieved from the following website 11/8/2009:.[9] Nkwa also includes the enjoyment of ahonyade. to restore life/vitality. Dr. perpetuate sangala (name)" Meaning to cure. to perpetuate" sanx "make live. He states: As one critically examines the prayers of the Akan in the traditional religious setting. one cannot help but come to the conclusion that the overriding concern is the enjoyment of nkwa (life). vitality/strength joy. This is not life in abstraction but rather life in its concrete and fullest manifestations. having sex (the act which brings life).pctii. vitality/strength joy. endure”) to recover (health. carve. sculpt. the calm one) sanx "bestow life. the calm one). have the same parent. make sculpture”) [Jesus was a carpenter] Meet. revive (dead). -sàngalaja preserve. vigour.org/cyberj/cyberj10/larbi. nourish. to have his equal. But life in Africa is not simply a biological pulse and a heartbeat. Rev. together. To last. to cure sanx "to nourish" -sàngaja The cross is a sign of a master healer and sage who utilizes his or her wisdom to bring life and vitality to the community. continue (<nenga “to last. in other words. to perpetuate" sanx "to bestow life. prosperity). that is. (possessions.html 91 .Hieroglyphic Root anx in Egyptian and Tshiluba Meaning Tshiluba -sàngalaja sanx "to bestow life. it means life of happiness and felicity. Emmanuel Kingsley Larbi in his essay The Nature of Continuity and Discontinuity of Ghanaian Pentecostal Concept of Salvation in African Cosmology69 supports this expanded African concept of living in the Akan tradition. These healer/priests among the Bantu are known as Sangoma. It means the enjoyment of long life. sculpture" Busongi Sculpture (<songa “sharp.t "throne" Sangoma? All kings are priests in Africa. return health to cure. and health. feed.t mAa. Sangila sanx "to bestow life. return health to recover (health. vitality. join. zambusu]. a true knower. below to mean the following: NGANGA ankh = life personified. and said: “If you season the policy of people and the community correctly. holder of the country’s equilibrium. for instance. doer. you are deified” [Watwisa mungwa ye nungu mu kinzozi kia n’kangu ye kanda. healer. This Kongo proverb shows us that only obedience to the people’s will makes people heroes and gods and not otherwise for the red carpet is not requested. and life free from perturbation. the name of a god reinterpretation = a master. teacher. priest. a simbi kia nsi. a specialist.” We mentioned that an Nganga is someone who unites heaven and earth: they are the mediators who “bind” the realm of the living with the realm of the spirits. a savior.” In Kikongo the phrase N’kingu mia zingu ye moyo means “the principles of life and vitality. Nkwa also embodies asomdwei. and substance. ba anx “living soul. riches. a life of peace and tranquility. spit on them. we therefore reinterpret the word anx. a title of Osiris at Tet” to be in Tshiluba BwaAnga “one who heals and protects: a hero. Coming back from an important and successful mission for my community.[11] (emphasis mine) Given all that has been stated thus far. in relation to the person. including the motifs adopted by other cultures like what we see among the Hebrew people of Palestine. one begins to possess the conceptual tools to properly interpret Egyptian phenomena. literally. it is earned [nkwal’a luzitu ka yilombwanga ko].” The term mianzingila embeds the Egyptian word anx “life. With this said. Kongo culture helps us to explain many of the concepts which are obscure in mainstream Egyptology.[10] including children. Fu-Kiau (2001:78) instructs us concerning a responsible leader in that: Political and diplomatic missions were akin to deification for those who knew how to handle the people’s responsibility. we reinterpret one of the titles for Wsr. This uniting and tying of knots is captured in the anx symbol which is a knotted rope. By studying the greater Cyena-Ntu (Ntu family) and its philosophies and cultural practices. a wise man took my hands. Ankh / anx Sn Snw 92 . that is.” The word anx is also in the Kikongo term n’singa which is part of a phrase n’singa-dikanda “the biogenetic rope of the family or clan. a power figure Budge 125A In regards to the nganga Dr.” N’kingu mianzingila is another phrase that means “principles of life.wealth. The reason why Jesus is associated with all of the attributes of the African cross (the anx) is because. Pygmy figure. We can posit as well that it is a symbol of “eternal life” as this is a symbol for eternity. Notice the similarities between the figures and the objects that adorn them. and some of them made great contributions to the kingdom.The Sn symbols to the right are symbols for “eternity” and “protection” in Egyptian. ka. If this is true. Again. The Snw on the far right is an elongated Sn sign and is used to enclose the names of royalty in Egyptian reliefs.” Below are art objects that span from ancient Egypt all the way to Nigeria. What we are seeing here is another sign for a “savior” as a savior is someone who protects: in this case it is the king who protects the country.a child's stool from the Cameroon 93 .Court dwarf's were occasionally attributed with supernatural abilities. ADDENDUM The Kongo may again provide us with insight into the use of ropes in connection with the anx symbol in Egypt and its meaning of “life. the Kongo provides us with the cultural framework in which to make a more informed analysis. the cross in Africa has a rich set of meanings that are all associated with the ideal human being in African communities of wisdom: the nganga (anga. This s-n root is cognate with our s-r/s-l root mentioned throughout this text. molded after the personification of all these ideals in the person of Wsr in ancient Egypt: Egypt’s founder and first king. In the next chapter we will see just what a real messiah is and how all of the terms mentioned above with the anx root help us to properly define a messiah. he was patterned. ga. As we can see here. We go through this in-depth so that we have a firm foundation to explore the messianic nature of Yeshua’. kwa. Benin . we can definitely see the Egyptian influence in his character. nkwa). The Sn is just an anx symbol without the vertical leg extension. It has been rumored that Jesus probably spent his formative years in Egypt as there is a full 18 years missing from his narrative. again. Dr. rope around waste and hands to the side or on the hips. rope around neck.Son of the Yoruba God Obatala. Fu-Kiua in his work Mbongi: An African Political Institution informs us of the reason young people wear ropes around their waists as can be seen in the figure from Cameroon. He goes on to inform us that: 94 . The Kongo may provide some answers to its initial meaning which we can relate to “life” and the anx. What we want to focus on right now is the ropes that are around the waist of a few of the figures above. From Ile Ife. Nigeria dated between 1000-1500AD Ancient Egyptian God Bes Nkondi PAKALALA pose – Kongo We see a common theme above that consists of: fat cheeks. skull pendant. protruding mouth or tongue. It is my contention that the anx is inspired from the thorax bones of the human body which unites the rib cage and the neck: both a set of bones that protect the body from harm (our internal savior). vertical one. a sign that the child was healthy and had a normal growth. 2007: 78-79) (emphasis mine) Here we are reintroduced to some common themes: health. We could then interpret the was symbol in the center as the brain stem. vitality/strength joy. to braid/weave") around its waist and neck (ciLuba nshinga "Neck"). the mother or the nurse could make sure that the child is growing normally in physical development (weight and height) and control of health. knot. When the circle of ropes become “Mpolunga (saggy) the mother was sure that her child was losing “volume” (Vonga/Vimbu) or weight (zitu/kilu). With them. These two ropes constitute “Mantantulu” if they are united by a third vertical [rope]. then the mother had to enlarge the size of [the] rope’s circles (zezisa/soba teso bia n’singa). At the center of each of these concepts is the anx. The “arms” are the shoulder blades. This is where I think the ultimate shape of the anx came from which later was associated with various other things (like ropes being tied around one’s waist which would give us the anx shape as well). tie with a wire. Only by feeding the child (or giving curative medicine in case of disease) can the rope “tighten” once more (ciLuba sangala "to recover: health. ropes and growth. Thorax bones in the center with “loop” surrounding the spine.70 The brain stem is connected to your spinal cord. encircle. braided rope". Another association with the anx that I have observed comes from the human body itself. it was a good sign. reel.According to African traditions a child wears a rope around his waist or N’sing’a luketo and another around his neck (n’sing’a laka). The health and vitality of the child is measured by the size of the “rope” (Kikongo n’singa’a “rope”. This was a sure “Dimbu” (symptom) that something was wrong with the child’s health. (Fu-Kiau. starvation and possibly the slow loss of life (vitality). On the contrary. 70 95 . when ropes become tight against the body (kala nakekete). the calm one”). ciLuba muJinga "Cord. digestion and heart rate. This interpretation can be confirmed when one examines one of the renderings of the anx that unites the spine Dd and the anx. A saggy rope indicates disease. roll up. shinga/Singa/Jinga "To bind. Sometimes these two ropes were united by a third. It’s responsible for controlling many involuntary functions. such as breathing. on the back. They are word during the period of birth until the age of three and were used as instruments of measure. thread. Ancient Egyptian fusion of the anx and the Dd pillar which is the spine. I can imagine that the need to understand this aspect of the body arose in Egypt as a result of a lot of injuries do to heavy lifting building temples and pyramids.We talked about Esu. I contend therefore that many of the “gods” were used simply to teach anatomy. which means it must balance a lot of weight. junction point of two shoulders or two arms= nTangani/diTung (U. we begin to appreciate a deeper aspect to the teachings. sought to understand and know life in every which way they knew how. in African centers of wisdom. all of the “gods” are associated with some part of the body or with an herbal plant. When we understand the relationship between the anx and the cross in relationship to the body. Among the Yoruba. junction. As a result of the scientific study of life. the thorax bone. The anx in many forms Back of an Akuaba among the Akan with the Dd pillar found among the ancient Egyptians Akwa-ba doll from Ghana 96 . understanding how to “restore health” in terms of bone repair was critical in ancient times. node" The thorax bones is the “crossroads” between heaven (the brain) and earth (the body) in terms of our physical anatomy. The burden of the spine and shoulders hits home for folks who have to move a lot of heavy things. As we can see with these examples. as well as future doctors (nganga) that were going to treat injuries to the different parts of the anx. Nshinga ma-Sang(u. Wsr and Yeshua being associated with the “burdens” of the world. joint.a-mapa/makaya/mapwapwa" "crossroads. Eṣu would be the sympathetic nerve system which is attached to our brain stem above. the ancient Africans needed a way to practically integrate the wisdom gained from careful study into the lives of the common person. Therefore. the cross (the anx) is associated with many aspects of “life” and the ancient Africans. in terms of bones. In all of the African spiritual systems I have had the privilege of studying.o) di-/ma-Sanga "Neck" "Interval between the shoulders. we find the anx in various forms across Africa: all reminding us of different aspects of life (physical and spiritual). A) . confluence. The anx is part of the shoulder and is connected to the spine. With that said. With Eṣu being the “crossroads” he definitely would be. Together the shoulder and the spine carries the “burden” of keeping the body vertical. Aduno Kine Dogon symbol for the “life of the world” From Lemba-Lakkous. Cyprus Asante akua'ba doll 97 . anointed.” 98 . we must properly understand what a messiah is. smear or coat” something. Anointed Language Greek: Χριστός (Khristos) Latin: Christus. The title occurs in the Hebrew Bible. They understood the basics. simply. Christ. To get a greater understanding of what a messiah is and his/her role within society. also to paint:--anoint. grease. is to “rub. sgnn "ointment". to rub with oil or fat”. What allegedly made Yeshua so special is that he was the Son of God which put him in a special category: as if to be the last Christ on earth. but could not progress further because they lacked the institutions that make messiahs that Africa has to this day. by implication. where it signifies the installation of a "king". weaken". sgnn "anoint. paint. anoint” “to expand. anointed. to anoint. is a title from the Greek language. the Biblical text and Hebrew culture in general is lacking. wrh m mrht “to anoint with oil. gs "anoint" (someone with). mashiach.” We also have a form related to our s-r root: n-sr "anoint" (injury). Egyptian and Tshiluba languages.” Budge translates urhu as “anointed ones. wipe. and/or military authority. chosen by God or descended from a person chosen by God. i. to consecrate. smear. There are many “Christs” in the Bible: especially in the Old Testament. mrh.” In Egyptian we also have wrh “to rub with oil or salve. Before we can engage this debate. sgnn "to rub". Below are the words for anointing in the Greek.” Another term in Egyptian is mrh “to anoint. obliterate". or moshiach Aramaic: mshikha s-kh s-h s-# š-g š-g ^^ ^^ “anoint. The word that many people mistake as a last name for Yeshua. This is primarily because they adopted the concept of messiah from people like the Egyptians.t “oil.” In the Egyptian we have a few reduced and modified forms: Egyptian sin "rub in. to rub with oil. advisory.THE MESSIAH Much debate enrages over whether Yeshua’ is THE messiah for all mankind. but did not adopt all of the vocabulary and philosophy that goes along with it. The “anointing” act was a part of the ritual for the installation of the leader. pat. fat of any kind”. clean. everyone is expected to develop into a messiah: so there was no need to “prophecy” about a messiah and savior. s-kh (ַ)מׁשִׁ יח ָ mashiah. soften. Kristos. coat or spread. unguent. to smear.” The word anoint simply means to “rub. massage. which means “anoint. smear. In Yorùbá we have ṣan “smear. Whereas in Africa. anointed” ^^ a primitive root. Hebrew: mashiakh Root Meaning “anoint. wash. suet. "prophet". to anoint. moshiah. rub out. Hebrew. smear on (unguent) [metathesis s-g > g-s). religious. a messiah was something unique among Hebrews because the concept was foreign to them.e. or "high priest": a person. Therefore. dip” “brush. to serve as a civil. to spread” “coating” (mpemba “white clay”) ()מְ ׁשִׁ יחָא English: messiah Egyptian: msw Tshiluba: shinga Langula Ci-shingampèmbà To anoint. The root of the Hebrew mashiakh is shiakh. bought spices. It should be noted that the Greek word Kristos derives from the Egyptian qrst “mummy. so that they might go and anoint Jesus after his resurrection. whoever wrote the story of Jesus was very familiar with Egyptian (Kongo) teachings and initiations and it is apparent from the evidence presented throughout this book. AFTER one has successfully completed one’s educational training in a life principles institute.(š-kh) Other terms reinforce this (via Strong dictionary): nacak (naw-sak') “a primitive root. and the qrst. especially a libation. Therefore it is no coincidence that we note in Mark 15:46. to smear over (with oil). melt. This was often done with spices. to bandage. there is no evidence he even existed. 2:39-40). or to cast (metal). However. let alone travelled to Egypt. set (up). was made in the process of preparation by purifying. 99 . Many speculate he went back to Egypt. to knot.A true “Christian” would be named a Mashiakh. i. This is why for them the simple act of ritually rubbing a priest or king with oil is enough to bestow the title of mashiakh on them. It is mentioned in the Bible that Jesus at an early age escaped an assassination attempt and went into Egypt (Matthew 2:13-23) where he lived for a time before returning back to Nazareth (Matthew 2:23. (cause to) pour (out). offer. Lk. Based on the linguistic characteristics and concepts associated with Jesus. X at all. and embalming. anointing. while others suspect Greece and India. these concepts (in reference to leadership) among the Hebrews are borrowed into the culture and they do not fully express the range of concepts that are concentrated together for this sk root (trying hard not to “seem” like other nations).e. embalmment. by analogy. 16:1 that Mary Magdalene. anoint:--anoint (self).” cuwk (sook) “a primitive root. But from age 12-30 there is no record of his life. I would vouch for an Egyptian travel.” The word krs/qrs denotes the embalmment of the mummy. This is not the case in Africa. as the mummy. At age 12 he is reported to have went to Jerusalem to celebrate Passover (Luke 2:41-50).” However. and Mary mother of James and Salome. properly. to anoint a king:--cover. One receives this title after demonstrating proficiency in certain skills and has demonstrated a high level of moral character. One thing is for sure. to pour out. com/2009/08/01/in-africa-18/ 100 . Massai warriors wear red body paint and often women are painted white before a wedding ceremony. Initiation is a holistic experience. all candidates were painted with the tukula (red) color which is a symbol of both danger and death.” the root of the word mashiakh is the word in Egyptian known as anx (Coptic onkh. As George T. we again have to leave Israel and go back into Africa. To be initiated is to acquire the highest knowledge that the community experience has accumulated through time and space. To be initiated is to be ready to accept responsibility (carry one’s cross). in an attempt as not to openly adopt “pagan” practices. Kikongo n-ga. abandoned the practice of full body “anointing” with clay and paint and therefore substituted these materials with oil: usually rubbed on the forehead. The ancient Hebrews. kwa. It’s all part of their initiation into manhood71 To be “anointed” meant to go through initiation. both as a mature human being and as a powerful spiritual being (what a cross. The color is a reminder to the candidate that this is a serious matter and helps to adjust his mind for the endeavors ahead. As in the case with the “cross. Literacy is more than reading or writing as understood in the Western sense. The whole concept of “anointing” describes a process in African initiations where the young initiate is “rubbed” and “covered” with white clay or red paint. maturity and leadership (Fu-Kiau. In the heart of central Africa’s Ituri forest. Initiation is the method of participatory research that emphasizes the importance of symbolic literacy in understanding pre-Western modes of thought and being (Kajangu. angle for their lunch. for this is the place where messiahs are born. Childs notes in regards to this phenomenon. boys from the Mbuti. anx represents). 2001: 128). 2005: 70). For instance. Body painting wasn’t only used for initiation.To fully understand what a mashiakh is. Duala ong. 71. To be initiated is synonymous with going to college or a university. Ftn 8). one of several pygmy [Batwa] groups in the area. For instance. Akan nkwa. “Traditionally minded Xhosa daubed themselves with ocher clay during initiation ceremonies” (Childs. which in a host of ways enables human beings to understand their destinies and to discover ways and means to fulfill them. This may be the origin of the white dress ritual common in modern times. which I guess you could consider these initiatic moments in one’s life. Tshiluba anga). 2003: 9. secret and sacred. in the Kongo to illustrate that the teachings were highly powerful. By double sight we mean the ability to use all the gross and subtle faculties of human consciousness to perceive reality in its totality—visible and invisible. A common expression among Bakala (African-American) people is that.” Initiation. they do not see. shiakh) is one of its rituals—that facilitates this process of self-discovery (of one’s self-healing power). A few African proverbs will help us better understand the value of the elders in African communities of memory: Umbundu Yorùbá A community without elders does not prosper. It is these wise men and women who are the “saviors” of the nation. To be sensitive to these waves is to be able to positively or negatively react to these forces. African education encourages you to discover. The ability to touch. Education in indigenous societies aim at pushing the boundaries of what human beings think is possible. therefore. it is “natural” to be part of nature and to participate in a wider understanding of reality (Somé 1994: 226). One aphorism that I have coined in trying to articulate the difference between Western religions and indigenous African education. destabilization of the body’s habit of being bound to one plane of being. Traditional education consists of three parts: enlargement of one’s ability to see. Those familiar with the story of Jesus can recall the famous verse in Matthew 13:13 “This is why I speak to them in parables: Though seeing. An elder cannot be present in the market and yet permit a child’s head to lie 101 . One is either sensitive or immune to these forces. Enlarging one’s vision and abilities has nothing supernatural about it. see.” Initiation is a process which helps to sensitize one to these waves and radiations so that one can “see” using previously closed senses: to better interpret the messages in the waves (the vibes). who are able to help the community solve its most complex problems. literacy is often referred to as “double sight” (de Rosny 1985). initiation helps to develop sages. hear and taste in the ordinary sense does not allow us to do these things in an initiatic sense. To have open senses (both physical and spiritual. Therefore. within the confines of a wisdom center. and the ability to voyage transdimentionally and return. In the language of initiation. one must know life to the highest capacity and it is initiation—for which anointing (shunga. 2) tower over the one thousand and one challenges of life and 3) to nurture greatness in all things. in my opinion. smell. It is about learning how to see the past to predict the future.” or “I don’t like the vibe in this place. is a process in which a person learns how to tie and untie “knots” (kolo): to code and decode ancestral wisdom (hidden in parables). “I got bad vibes from her. they do not hear or understand. literacy holds a very expansive meaning. We live in a vast sea of waves and radiations which carry all kinds of messages from many sources. Malidoma Somé in his work Of Water and the Spirit provides the best explanation. It is feeling and holding tightly the past segment of the biogenetic rope in his/her hands to insure its linkage to the future.For African people who have been initiated. as to the purpose of initiation (education) in the African schools of wisdom.” What one is encouraged to discover is the true nature of the universe and how to use the wisdom of nature to 1) make life more beautiful on earth. In order to fulfill these lofty goals. is that: “Western religions require you to believe. though hearing. Wise elders are the community’s immune system against the syndrome of foolishness that plagues human societies. rather. It is the elders who protect the community. things go away. symbolized by the D) means to walk on the path of mastering the knowledge of life. An elder’s face never gets wet by the rain. save" (<sunga) [l+i>di] "separate from. The youth walks faster than the elder. defending against. make happy. appease. community” [usually at a community mbongi. Without an elder a village is unhappy. What an elder sees sitting down. continue” (<nenga “to last. When an elder speaks. help" "separating (fighters). persevere” “to cure. continue. endures. return health” “Love. he was born [for] long distances. boka] 102 . communion. One does not go empty-handed to see of an elder. You may refuse the elder the meat. A human being becomes kind in the evening of his life. Know the limits of merry-making because the elders are like mountains.Akan Shona Akan Ntumu Swahili Twi Asante Ntumu Mossi Nembe Lega Luo Oromo Ovambo Wolof crooked.) He who does not listen to the advice of an elder will see bad things. As one gets older.) are the embodiment of wisdom and it is wisdom that staves off foolishness which can bring catastrophic disharmonies to a society. success. meet(ing). conciliate. Childhood comes twice. a youth cannot see standing up. An elder is a big basket that gives protection from the wind. [but] you shall give him the liver. the calm one) “To last. have joy. vitality/strength joy. An elder engages no idle dalliance. reunion. value. charity = diswa “Sustaining” (sustain) “love each other” “To please. but the elder sitting on the ground. The child looks everywhere and often sees nothing. defend against. you do not abandon it to speak childish language. protects from the wind. An elder’s fart does not smell bad. The rich person is the elder. rejoice. If you keep on playing with the elder’s tail. issue. one grows tall in wisdom. happiness. Hrw. persist. to be delighted” [know how to do an] “Invocation ritual” [has] “Price. protect. The mouth of the elder is more powerful than a charm. The strength of the elder is in the ears and on the lips. pacify" “to recover” (health. expand. it is because he has something worthwhile to say (advise. rejoice” “Be happy. A powerful man. The elder [is] a turtle. like the big basket. joy” [initiate a] “Gathering. Ogboni. you will end up in a sad mood. dignity = bunema” “To welcome” (to embrace someone) [is the embodiment of] “good fortune. A difference in age shows in a difference in experience. If the elders leave you a legacy of dignified language. but the elder knows the road. sees everything. endure”) “Durer. The Nganga are those who know how to: sùngila sungidila sunga sangala Nanakana Nenga sàngalaja Dinanga Nengesha Nanga-nanga Sankisha Sanka muJinga MuShinga Sangesha Mu-Sangelo Sangila "deliver. etc. The elders (Bukulu. The birth of every child can be said to be mushangi. the elders in training. a mushangi is a spirit who has shed his skin and come back. the initiated. the nganga (the elders. coming back In other words. reconciler (m-s-g) This s-g/m-s-g root is also associated with another term that is the cornerstone to the Christian faith: Shanga >muShangi/-Shani Undress. In Africa. but did not adopt all of the practices. This is done in order to try to make themselves appear unique and special in the eyes of others who may not be well educated on world history. the healers. being bodiless Spirit. one who has died and resurrected: a spirit returned. the leaders. we can see how much substance has been lost in the Hebrew’s modification of African cultural practices. the visual motifs and the lexical items associated with these practices. By not reduplicating the institutions associated with these terminologies. the Hebrews lost the essential essence that inspired the Jesus myth in the first place. comparative religious studies or linguistics. a “messiah” embodies a range of concepts that are lexically themed with an s-r and s-k/s-g (m-s-k) root. The modern Christian is in the habit of trying to monopolize certain terminology to make it seem as if these concepts and practices are exclusive to Christianity: to hide their “pagan” African roots. redeemer"(<sùngila) (m-s-g) Musungu “peace maker. the arbitrators) are to the community a: Hebrew shiakh “anoint” (s-k) mashiakh “anointed” (m-s-k) ciLuba shinga “anoint” (s-g).Therefore. By examining African languages and cultural practices. for which they adopted certain concepts. Ci-shinga-mpèmbà “coating” (initiated) musùngidi "savior. 103 . Just as there are many types of trees that serve various purposes in a forest. El. a Musungidi (the personification of wisdom). the hidden land). the place of dead souls (the grave. Asar) in ancient Egypt was Abjw (Abydos). in a community of Ngangas is what saves the community from the ills brought on by the spirit of conflict (Egyptian isft. In this name for the place where Wsr ruled is the word ju “mountain-wilderness. since the time of unrecorded time. is designed to educare “draw out” (educate) those gifts from the child and teach him or her how to best use their gifts to enhance the community in which they reside. As mentioned earlier. Kanga) serving in various capacities: reinvigorating its vitality. So for a muntu. There is no one thing that is the “end all. Kulu. Abju indicates the place in the wilderness of the mountain desert on both sides of the Nile. Ju means “mountains” in Egyptian. So for the BaMelela (the African). Gueno. the city of Wsr (Abju) is cognate with Yorùbá iboji. These places of initiation often take place in the forest or the “bush” for people who live in the plains areas. 104 .” It should be noted that Yeshua also went down into Sheol after his death according to Ephesians 4:8-10 and 1 Peter 3:18-20. We also have in Egyptian sr “to reach down. the Africans institutionalized as part of the general education system in an effort to reduplicate as many “messiahs” as humanly possible within their communities. the BaNtu. To develop this legion of wise community pillars. This would make God (Godo. be all” in nature. have opted to create institutions of life that increase the likelihood that the person has the ability to contribute to society at an earlier age. The principle cult center for Wsr (Osiris. Amazulu uzibuthe) and the syndrome of foolishness.” In the Yorùbá language we have two cognates to the name Abjw which are both heteronyms – totemic variations on the same phonemic root: iboji iboji “shade” (ib-ojiji “the place of a shadow. thus why the name of Wsr is reflected in the Yorùbá word isale “down below” (s-r/s-l root). an Nganga. We know Wsr is the lord of the underground (Amen-ta. everyone born to a community is born with a mission to serve: to “save” it from destruction and chaos.” In Hebrew the name becomes Sheol (š-l) “the place to which all dead go. iboji). therefore. I say this to say that communities of memory do not wait thousands of years for a benevolent spirit to return to earth in order to bring balance back to society (to save them). We know the Creator works with efficiency because we study nature (the nTrw) which is a system of systems that works at maximum efficiency. It is the forest or the bush that is considered the “wilderness” and it is where Yeshua went through an initiatic experience in the Biblical text. shaded place” “the grave” A shadow evokes images of darkness as a shadow is black and obscure. Africans. the way the Hebrews have interpreted a messiah is incompatible with the efficiency model espoused by African people. In other words. Allah. The community of memory.THE AFRICAN SCHOOLS OF LIFE What the Hebrews tried to embody in one man named Yeshua/Yahushuwa. Eloah) very inefficient and ineffective. The spirit incarnated as a child in the womb is a package full of gifts to be shared with the community. All things are interdependent on each other and all things (ntu) are embedded with certain “gifts” that are used to serve the greater good of adding life to life. they do not leave things to chance. it is the muntu’s job to be a “messiah” as a messiah. In Africa. so too are there various “messiahs” in the forest of community (the Kongo. Hrw. Oluwa. also from the Kongo. how to transfer strong and healthy life (Fu-Kiau. Yorùbá iju “jungle” is cognate with Egyptian ju “mountain. oganjo oru. how to PROTECT it. ibid. redeemer"(<sùngila) are trained and developed. there are institutions for which “messiahs. The farmer didn’t have a desert to contend with like in the Sahara. These builders of symbolic worlds manage the destinies of African wisdom traditions. and above all. But one cannot farm amongst trees.” The wilderness for the Yorùbá was the jungle: for the Egyptians it was the inhospitable desert of the red mountains (dsrt). with practically the same name.” Ngangas. “the institute that folds/ties and unfolds/unties principles of knowledge) Ku Londe – The High “Site” (of knowledge in the physical world) Ku Kongo – lit. in his unpublished PhD dissertation title Beyond the Colonial Gaze: Reconstructing African Wisdom Traditions (2005) provides insight into the character of an elder who in our case would be the nganga (anx) or which is sometimes called in the Bantu the Bakoles or Bakulu. It is this reservoir of knowledge that enables elders to be effective teachers. The root anx can be seen in a few of the names for these institutions. “present/yes”.” the dead of night. First. Institute of the called ones Sansulu bia Zingu – Institutes for Life Budulu bia Meso – Eye Enlightening Institutes Ghandusulu – Initiation or Training Institutes Tambukusulu bia N’kingu mia Zingu – Life Principles Transferring Institute These institutions (Sikudukusulu) were created to serve the muntu (the person). with the sole intent of creating Ngangas: Pala/Pila dia Zingu pr ankh = house of life (the name of a college of priests) reinterpretation = wisdom center for the development of masters (priests) Budge 124B Dr. The elders or living libraries in a community of memory are people who are: … [M]otivated actors who construct symbolic worlds in which people live and die in Africa.In Yorùbá iju means “jungle” to the rain forests dwellers. These sages exhibit the following characteristics. The inhospitable wilderness of the jungle is described in Yorùbá as aginju “dark wilderness. soil) is from the same root oganjo (ogan-jo) “the dark part of the day. Second. a musùngidi "savior. they 72 Oduyoye (1996:119) 105 . fertile. Kykosa Kajangu. egàn “the soil under the forest cover” – black.” The underlying semantic spirit of each term demonstrates a primary connotation of “wilderness. Fu-Kiau (2003: 2) provides us with the names of these cultural institutions in the Kongo: Ku Kanga – site of initiation (lit. they have learned to the highest degree the secrets for knowing life and the strategies for stemming the tide of its challenges. they were devoted to assisting the human being in understanding the key principles of life: How to LIVE. These very institutions were present in ancient Egypt.72 In the Kongo.).(cf.” The qualifier agin. how it is UNDERSTOOD. elders. In regards to the nganga Dr. that is the anx just isn’t life itself. holder of the country’s equilibrium. They have developed countless strategies or teachable viewpoints to take people to places where they have never dared to go. and these ideas conceptually that are embedded in the Hebrew term mashiakh (messiah. but the obtainment of the highest degree of the knowledge of life and the strategies necessary to tower over the one thousand and one of its challenges. ancestors and the divine aspects of nature. after he or she has completed their education in the schools of life mentioned earlier. Third. you are deified” [Watwisa mungwa ye nungu mu kinzozi kia n’kangu ye kanda. it is earned [nkwal’a luzitu ka yilombwanga ko].73 Without this initiatic object. a simbi kia nsi. who knows life to the highest degree and how to successfully meet its challenges that are worthy of the name nganga and can truly say that they have in fact lived in every sense of the word. Kajangu adds an important element to this discussion that I think we can add to our expanding definition of anx. zambusu]. a wise man took my hands. If a community member was in fact a doer (nganga-a performer of service to the community) then he would be deified among the people. He was the keeper of “knotted information.e. (emphasis mine) Dr.” In the past one could see an individual in the Mbongi who was a member of the community named “Na Makolo” (or Makolo in short). they have proven track records of success in taking direct responsibility for the development of the youth in their community of memory. spit on them. But it simply refers to the matter at hand. The nganga is seen as a therapist who is invited by various villages to deal with any issue the community may be facing. symbolically written 73 Mambu is what became known in the west as mumbo jumbo. Fu-Kiau wouldn’t find any disagreement with Dr. Kajangu’s assessment of the role and characteristics of an elder (the community masters. Before we continue. Remember that the term “god” commonly rendered in the Egyptological literature actually refers to initiated priests. Coming back from an important and successful mission for my community. the anointed) which they are unaware because they abandoned the culture and spiritual systems of their forefathers.embody the teachings of centers of wisdom. a specialist who deals with social issues. and said: “If you season the policy of people and the community correctly. Fu-Kiau (2003: 30) informs us of this practice and the titles of those who have mastered the art of “tying knotty ropes. the community will not accept him as trained or qualified. the nganga must present to the people his nkondi (his diploma) to ensure the village that he is qualified to discuss the mambu. Nkisi Nkondi An nkondi is like a diploma given to an nganga. the nganga). This individual was very important to the community.” i. Dr. This Kongo proverb shows us that only obedience to the people’s will makes people heroes and gods and not otherwise for the red carpet is not requested. It is these characteristics that make the living worthy of deification after death by the community. 106 . It is from this root linguistically. Before any discussion of the problem to be solved. literally. It is the learned and skilled sage who can bring balance to the country. Fu-Kiau (2001:78) informs us as to what makes a responsible leader and why they are deified after death: Political and diplomatic missions were akin to deification for those who knew how to handle the people’s responsibility. I would like to discuss briefly the ancient practice of “coding” and “decoding” knotty ropes which is essential in understanding what is about to be revealed concerning the Nkisi Nkondi. weeks. the root being anx. The Na Makolo was obliged to know in detail the meaning symbolized be each knot which he ties or each cut that he makes. This specialist ties and unties the Mambu or dealings inside as well as outside of the community. If you see an nkondi figure with mirrors as eyes. The nkondi was truly a document on which contracts were signed. or months.org/full-image. Deciphering these knotty messages was also the job of the Na Makolo. it is symbolic of written words to tell you what the nganga does—that he is trained to see what goes on in the actual society: he has exceptional insight (Egyptian sia). the person(s) who signs that document has to act upon that contract or that agreement in his life. He alone can deal with this. Any time the Mbongi made a decision related to numbers of days. in accordance with the decision made in Mbongi. the nganga will tie a knot with a rope and then they (the persons in conflict) will nail it on the nkondi. the specialists in “problem hammering” among the Kongo. It means someone is willing to make a decision.aspx?page=1288&image=kongo-nailfigure 74 107 . “archivist” in today’s Mbongi. 1880-192074 Nkondi (Kongo) with crossroads pose Image taken from. Ne Makolo was asked to represent that “Mandaka/Ngwizani” (contract) in cutting marks (Makenko) on a piece of wood made for the purpose. This practice was also in use among the Nganga-Nkondi. The nganga who carries this nkondi tells the community that he has the knowledge to deal with such situations. The nganga is an arbitrator and will lead a person to a resolution. and the word anx in Egyptian can also mean “mirror. After they have come to a resolution. through ceremonies and through medicines (nkisi). The knots can be tied by an individual or the community. From that time on. Na Makolo. Kongo nail figure used to fix oaths and heal the sick Democratic Republic of the Congo.information.” The nganga can deal with issues that link the living and the dead. set price or dates. If the Mbongi had also concluded a contract of alliance with another Mbongi.wellcomecollection. the Mbongi asked this individual. This information-keeper individual could be called “secretary” or better. to braid a cord and tie the number of knots which. would represent the date set up for the next debate or meeting (fig. 3). This is interesting to note because the nkondi represents the nganga himself. This means that all of the community’s “burdens” are presented to the nganga for arbitration. 108 .” As if this wasn’t enough to cast doubts on the Christian’s claim of originality of these concepts. which means he operates in both worlds: he is the knot in the biogenetic rope between the two communities (ancestral and living). the nganga.” It is no coincidence either that the word D (the cross) in the ancient Egyptian language means “oath. 3) the nkondi figure is used as an instrument to make contracts and record history by nailing knotty ropes: it is an oath instrument and binds a person to his words. My second interpretation is that Yeshua (if he existed) was a Kongo priest and to take one’s “cross” is.It is vitally important to remember that: 1) the full term is Nganga-Nkondi. Yeshua. to take one’s nkondi (one’s proof of authenticity) and be the arbitrator they trained to be. Wsr—is associated with the cross and the crossroads and each are brought to bear the world’s burdens. treaty or pledge. It is no coincidence that the Nkondi is a “wooden” instrument used by the priest to legally bind contracts with nails (also used in medical remedies: nkisi) and that Yeshua was “nailed” to the cross (the anx) which ushered in a new “contract” between God and humanity. swear." (Mark 8. Remember.” Even modern Christians interpret the cross as a symbol for burdens. as mentioned before. what we are looking at is another version of the “cross” in Africa and it is rightly dubbed an anx D: an nganga (nga reduplicated). So now we better understand what Yeshua means when he stated that: "Whosoever will come after me. and 4) the nkondi pictured are in the “crossroads” pose to signify point number 3. and follow me.34. agreement. as it is on the cross that Yeshua carried the burdens of the world through his “stripes.” Each figure—Eṣu. vow. promise. Remember that the anx is often depicted as a “knotty rope. in essence. the nganga are invited to various villages to be arbitrators and healers.) Yeshua made this statement long before he was “crucified” on the cross. an arbitrator for the community. is also a mediator. Yeshua and the disciples followed the same milieu. let him deny himself. With that said. The first is that this statement could be speaking metaphorically in terms of equating the cross with “burdens. and take up his cross. So what could he possibly mean when he says to carry one’s cross? I have two interpretations. Instead of using the word “contract” most Christians use the word “covenant” which is a “contract.” This is critical to understand because this may answer why Yeshua had to be “nailed” to the cross. 2) the Nganga has the insight to deal with the issues of the living and the dead. Connections like these are only possible when we unite Christianity with its source(s) in Africa. This term is cognate with Egyptian SA “to read. Lagos. to ordain. Eṣu.root of power Hebrew To come to pass swh Yorùbá se Mende Tiv Nupe Chu-chewa To command siwwah To cause to come to pass A command. The root of their names. authority The power to cause to happen Saw (imp.t “something decreed. 19:16). enjoinment. to design.” SA. to begin. discipline. In other words.” In Africa. 109 . In Coptic Ast is pronounced Ese which is close to the Yorùbá àṣẹ. cannon. was he a king because he actually lived and was revered as a king among the ancient Hebrews? Or is the title of King associated with Yeshua because part of his name belongs to a common root that the creators of the myth utilized to help facilitate his narrative? My vote is for the latter.KING OF KINGS Yeshua is known as the “King of Kings” throughout the Christian world (Rev. This root can be seen in the following languages (derived from the Afro-Asiatic swh root): The -s. when referring to their royal status. to allot. to decree. impost. Both Eṣu and Wsr in the Yorùbá and Egyptian traditions respectively are associated with royalty as either a prince and/or king. commandment. imprecation. it is the queen who gives legitimacy to the king. biding. command. to commission. ordained by God. derives from roots that deal with “commanding” and “authority. àṣẹ. dues. Àṣẹ: a coming to pass.t “the goddess of primeval matter. precept. Nigeria. The question now becomes. 17:14. virtue. power.s) Se (ekpe) mi-sewah ase ase tsav 75 Dictionary of Yorùbá Language. imposition. document. authority. taxes. to authorize.” SaA “the source of life. law. and Wsr) is how we would pronounce in Yorùbá. to determine. it is through the woman that a man becomes king. effect. (1913).” It is my contention that the root of each name (Yeshua. consequence.” SA. revenues.75 Àṣẹ is “the power to make things happen” and can be gleaned from analyzing all of the terms above. Church Missionary Society Bookshop. We have already mentioned before that Wsr and Ast are two sides to the same coin: the masculine and feminine to the same concept. instruction. “Eṣu represents the principle of life and individuality who combines male and female valences (Thompson. This concept is reflected in the ancient Egyptian language with the word hkA “words of power. ciLuba 76 The feminine aspect of Eṣu is Oshun.” It should be noted that all kings in Africa are priests. Here we have the visual representations of the two types of àṣẹ mentioned previously. spring. As Thompson notes. Egyptologists simply render the word As as “throne” but this is not an accurate reading. the source of life. Aset (js.” Duala hango “magician. Somali AySitu.magic witchcraft Ruler The angel who holds God‟s staff of authority Eṣu sawa tsav tsav Etsu Ma-sawe In ancient societies. The shepherd’s crook G in Egypt. The throne is a symbol for the “seat” of authority where the queen and king make their commands for the kingdom.t) Wsr Examining Aset’s name provides additional insight. The thronemrepresents the àṣẹ of political power. When examining the glyphs for names As. be the first. Eṣu also represents this principle of life (being a causal agent).t and Wsr. The word for the emblem of authority is the same word for the ability to command: words of power. so the equation of the shaman.t/As. This means the words for authority are synonymous with the power of speech.” SAa “to begin. procreative power (àṣẹ). Remember that the majority of the depictions of Wsr are with the shepherd’s crook. scepter. magic” (Coptic hako “magician. spells. queen or elders that were considered the law of the land. crosier. This provides us additional evidence of the crook’s meaning. 110 . we notice that they both utilize the throne glyph with the js/As sound value. The egg 9represents the biological. I think this strengthens our case for the non-mytholized Yorùbá concept of àṣẹ as being the linguistic and conceptual equivalent to the Egyptian goddess As. it was the “commands” of the king. the sign of kingship. the feminine suffix -t a represented by a loaf of bread. shaman”).t “crook. and the determinative of an egg9which is the symbol for motherhood (child in the womb) which carries the swḥ.t) Aset (js. words of power and kingship is warranted.t (Coptic Ese.t/As. originate”). is called hkA.76 We should note that in Kiswahili the name Eshe means “life” as well.t pronunciation matching our swh Afro-Asiatic root. 1984: 28). The first variation of Aset’s written name consists of a throne with the phonetic value of As/jsm. the egg is a symbol of new life to be (Egyptian SaA “to begin. ir. to promise.” We‟ve already connected Ast with the Yorùbá concept of Àṣẹ.” In terms of being a king. to prophesy" "proclamation" As we can see with this s-r root many concepts associated with Yeshua: “prophet. Robert Farris Thompson in Flash of the Spirit describes an art object depicting male and female Eṣu with bulging eyes. The vast majority of the depictions of Wsr’s name utilize the “eye” glyph to represent the r sound value. which we equate with Hebrew shuwa’ “riches. The j and s sounds interchange frequently in African languages: especially in Yorùbá and Tshiluba. but ir/ir. to announce. to initiate something. who has the authority to make things happen and have an effect for which these words of power are directed. Not only does this glyph mean “eye” (ir.t). we are reaffirmed that Eṣu is the personification of àṣẹ. desire” (<swa “love. construct. prepare. do. good luck and good fortune” that may be relevant to our discussion and related to Yorùbá àjé. Oṣun among the Yorùbá is also the goddess of good fortune and prosperity. to reveal. make. want”). will. DiSwa also means “love yourself. “maker” (of things) “create. act. In describing its meaning. who has the words of power to command (àṣẹ). a maker. someone who acts. 111 . Thompson notes that: [the eyes] embody the power-to-make-things-happen. Oshun is the goddess of “love” among the Yorùbá. achieve. magistrate" "official" "foretell. to challenge" "prophetize. Coincidence? What we are seeing in the rendering of Wsr’s name is a cryptographic/ideographic reading which states that the king is someone who “has exceptional insight. 77 Ast is also known as the goddess of divine love.” Although they are two different glyphs (the m and the ).t “doer” (of good). make known. make” All of these renderings are connected by the underlying idea of “making something happen” and the power to do so. beget. the gift Eshu received from God in heaven (Thompson. wealth. prophesy. treat” “swear an oath” “administer. 1984: 28). is a doer (nganga). with two different sound values. It should be noted that there is an ancient Egyptian god by the name of SAw “the god of prosperity. to make come to pass. With that said.” Underlying all of these different renderings for this -sroot is the “ability to make things happen”: to “cause” something to be.t (with the eye glyph) has the following values: ir/irw iri iri ir.t "nobleman. be self-satisfied. (emphasis mine) This is very interesting to note. be proud. This homophonic root is in the god Eṣu and Eṣu‟s mother is Oshun. Cross comparing this root in Yorùbá in the name Eṣu.DiSwa/CyAsa77). in Yorùbá we have the title Awujale “supreme king” (w-j-r). The awujale is a ruler of a kingdom or confederation in Nigeria. these concepts should transfer to anyone with authority and we can see this associated with our s-r root so essential to this conversation: sr sr sr sr sr. promise. manufacture. to spread abroad. Our ciLuba rendering of diswa means “love. their conceptual values are practically the same. ) (Campbell-Dunn 2009a. King of Kings. in regards to Eṣu and Wsr. THE KING OF ETERNITY. Thou art he whose White Crown is lofty. and the crown of thy head is of One can see E. etc. This is cognate with Egyptian D3D3 “head”.” ìwèjú “mystic vision. We point this out to say that Wsr exists in Yorubaland as a king.e. the word for “head” can be used for the features contained in or on the head (i. Thy existence endureth for an infinite number of double henti periods in thy name of "Un-Nefer. 1967 (<1895). appearance. face. Thou hast governed the Lands of Akert. Thou hast ruled the Two Lands from the womb of the goddess Nut. Thou drawest on that which hath not yet come into being in thy name of "Ta-her-sta-nef. Dover Publishing." Homage to thee. aperture.” The -j.root can be seen in such Yorùbá words as oje “lead” and aṣaju “leader.” We discussed earlier. thy head is of lapis-lazuli.” iwùjè “forehead. the word iwájú-orí means “power center in the forehead. THE GREAT GOD WHO DWELLETH IN ABTU. eyes. the Erpat. dje.” The word ori (Egyptian ir “eye”) means “head” and can also mean “consciousness. Coptic djo. O thou who art in the kingdom of the dead.” Within this same root we have iwájú “forehead. and the whip.A. 8-11. edge of a knife or sword. Thy members are of silver-gold. boundary. nose. he is the “king of kings” (an emperor). seat of mystic vision. WHO TRAVERSETH MILLIONS OF YEARS IN HIS EXISTENCE.” àwùjè “top of the head. and Prince of Princes. precinct” (related to ile “house. The king is literally the “face” of the nation..Wsr = Yorùbá Awujale “supreme king” Awujale contains the words ojú. and Lord of Lords. Wsr is also considered the nsw nswtyw “King of Kings. the concept of the “eye” being a symbol of exceptional insight and the power to make things happen. THE LORD OF EVERLASTINGNESS. Thy son Horus is firmly placed on thy throne. and as the Heq who dwelleth in Abydos. In the Ancient Egyptian tradition. awareness and destiny." Thy power is wide-spread. seat of mystic vision. Thou art the eldest son of the womb of Nut. Wallis Budge’s translation with hieroglyphs in the Egyptian Book of the Dead: The Papyrus of Ani. land”). iwaju “face” and alà “landmark. HYMN TO OSIRIS UN-NEFER 78 A HYMN OF PRAISE TO OSIRIS UN-NEFER.” Below is a hymn to Wsr which explicitly states these words. thou art he of whom the fear is great in thy name of "Usar" (or "Asar"). limit. Thou hast gained possession of the scepter of rule. In other words. 78 112 . Thou hast ascended thy throne as the Lord of Tetu. 2009b). pp.” In African languages. the “supreme king” over the entire confederacy. But not just any king. Thou art the King (Ati) of gods [and] men. This is another way to say the “head” and it is a metaphor for being out in the “front. Thou was begotten by Keb. Thou makest the Two Lands to flourish through Truth-speaking. front. confine. and the rank and dignity of thy divine fathers. Here Wsr (w-s-r) = Awujale (w-j-r). look." Thou governest the Two Lands by Maat in thy name of "Seker. In liturgical Ifa (the language of the priesthood of Ifa). lips. in the presence of him who is the Lord to the Uttermost Limit. Thou art the lord of the Urrt Crown.” The word ojú can mean “eye. Thy heart is expanded with joy. present. publish. Olu. In many places in Africa. Grant thou to me glory in heaven.turquoise. height.” So an alãru is “one who carries the burden”: a messenger. The first messenger of the Bible was Moses. carry. and journeys far distances to “extend” trade alliances with kingdoms. niombo. ms "bring. with wheat and barley therein. without obstruction. Mw-Ene. who has access to authority. ms "to launch" (an exodus?). bearer of burdens. that if anyone wants access to God. Wsr. Kings are divine in Africa and this ritual imitates what we observe in creation. Ilu. and [the power to] sail up the river to Abydos in the form of a Benu bird. Let there be given unto me bread-cakes in the House of Refreshing. age. spreads the news abroad. tell good. announces. nTrw) to convey its desires. preach. El. that upon further investigation. Thy body is all pervading. rod (of Osiris). who carries the message rod or staff of authority. to have access" (to the king? God?). extend.” The al. and [the power to] pass in through and to pass out from. and [the power to] sail down the river to Tetu in the form of a living Basoul. may prove to give the real meaning of the name Moshe (Moses): Egyptian ms. the official “reveals. Kulu. a possesser of. makes known” the desires of the king: he is a “messenger. The belief is God is too “distant” to be dealt with directly. ms "bring. and Prince of Princes (hq hqw). Allah): all the roots denoting “distance. see Credo Mutwa’s book Indaba My Children (1964). Therefore. bear. he/she is someone who acts on behalf of the king and relays the “commands” of the royal office. offer". God didn’t speak to the people directly. and sepulchral offerings of cakes and ale. Yeshua respectively). This is the idea behind the word “God” in the first place (*godo. God doesn’t directly speak to its creation: it uses its messengers (the òrìṣàs. and a permanent homestead in Sekhet-Aaru. cognate with oni which means “one who has.79 In Yorubaland they have what we call alãru “carrier. and truth-speaking in the Divine Underworld. Eloah. It is the royal linguist that relays the message to the community and the people speak to the king through the messenger. 79 80 113 . msw "carriers. the king doesn’t speak directly to the people. With that said. he spoke through Moses.” In Africa messages were carved into royal message sticks which were sent with messengers to other kingdoms to be read by the king in that locale.to the Double of the Osiris.” This is not a “commonality” but an exactitude and cannot be easily dismissed by Christians. bring. Yorùbá Ìbarà / afárá “bridge. according to the Bible. one who is the “bearer” of good or bad news.. journey". Egyptian and Christian traditions.” For these reasons and more. In terms of an official. also took with them “booty” from the Egyptian storehouses and treasury. the scribe Ani. travel. they must go through his messengers (Eṣu. ms "to be admitted. and propitiatory offerings in Anu.” The /b/ and /m/ sounds interchange in African languages (i.80 I think all of these terms can refer to one aspect or another of a messenger: one who “brings” something. bring away booty. Thou art An of millions of years. canal”). *gudu. it is attested that Wsr (Jsr). water channel. before Yeshua is alleged to exist among the Hebrews. bearers". porter. was called “King of Kings.” This is personified in Eṣu who is the “angel who holds God’s staff of authority. shew forth. Gueno. titings.” A “close” reading in Hebrew would be the word basar “messenger. present.t "staff. Lord of Lords (nb nbw). ms "to extend" (hand). we observe in the Yorùbá. You don’t hear of God speaking to all of the masses at once: he always uses a messenger (human or angelic).e. and power upon earth. msa "march. we note the following terms.is a prefix. travels. As we can see here. the doors of the lords of the Tuat. O Beautiful Face in Tatchesert. For many examples of this in Southern Africa. who marches. The Hebrews once they left Egypt. But let’s continue our conversation. water ford” Egyptian mr “bridge. take (aim). Moses could simply mean “messenger. opens the way” for business. the messenger is someone who “clears the path. commissioner. Not only do all three of our figures represent the ability to make things happen (the life principle). so while he is at once the king. emissery. agent. work. path. command. trade. bystander. As above. that in our respective areas of focus. It makes perfect sense in the grand scheme of things as the king was a representative of God on earth. an access point). he is also a messenger himself: the knot of the anx D that binds heaven and earth. communication." It may further break down into wp "door" (a path one enters. but they are all represent the “channel” or the “pathway” to God. so below. It is clear. 114 .It’s interesting also to note that one official term for “messenger” in the Egyptian language is wpwty/ipwty “messenger. matter. errand. job. and wAt “road. commission. way. matter. chore. This same process works in the spirit world as well. This leads us nicely into our next chapter. business. assignment.” It derives from the word wpwt "orders. expedition.” There is a deity by the name of wpwAwt “the opener of the way” who is associated with Anpu and Wsr (the messenger). side. etc. The important thing to note here is the connection between air. Assman further elaborates on the role of Shu in Egyptian theology as the life-giver: It is his duty to give life to creatures and provide them with the conditions necessary for life and keep them alive and he does this on behalf of his father Atum the creator god. Jesus said.” In other words. and the life. if you wanted to get to the Heavens from Earth.OSIRIS. ESU. (Assman. "I am the way. Yeshua and Wsr. tr. For instance. The question then becomes. Cornell University Press. 81 115 . Egypt and Kongo..” Ancient Egyptian SHU. Shu being the “breath of life” essentially plays the same role as the Holy Spirit. they no longer say Yahweh. we find deities and concepts that represent a “path” or “channel” that contains our s-r root. Egyptian Solar Religion in the New Kingdom. David Lorton." In other words. the Egyptian god Shu (Sw) refers to the “luminous air and light filled living space. One of Shu’s epithets is “life” and he is See Jan Assman. the pneuma “breath or wind” of God in Christian theology (Murdock 2009: 325). The word Sw has several meanings depending on the determinative attached to the root. is Yeshua the “way” because he actually existed and is the son of God? Or is Yeshua the “way” because he was borrowed from the African deities who were also the “way” based off of our s-r root? My research supports the latter. These concepts can be realized in a spiritual and physical sense. if you are looking for a way to get to the father. In our three pre-Western centers of wisdom of Yorubaland. in modern Yiddish Hebrew. symbolized by the sun. I wonder if there is any connection between the Indian god Shiva.”81 Shu is the air between the earth and the heavens and also the “breath of life. there is no connection of Eṣu of Yorubaland with the sun. SYUA. the truth. the Sun.. Yeshua is the road. Wsr is also a solar deity and we can see it written in his name below. Shu is also responsible for the air one breathes. sun and sunlight. you had to go through Shu: literally. 1995: 178) It is no coincidence that the word Shu is also contained in the names Eṣu. all being represented by the word Sw. the Sun To my knowledge. and our Egyptian god Shu (Shiwa).. but Eṣu’s core meaning of “power” and “life” may provide a conceptual connection. For example. One rendering of Sw can mean “dryness or emptiness. Often the /v/ sound derives from /w/. space. 1995 p. but Yahveh.” Some translate this as “void” and speaks to the area between heaven and earth. No one comes to the Father except through Me.81. This word can also mean “ascend. More investigation is needed here. JESUS – THE “WAY” In John 14:6. the sun and the light present that fills that space. daylight Ki-Kamba-Bantu SIUA. This image also speaks to that process. whom Atum fashioned. the Coffin Texts spell 464 mentions the “blood of Shu” who we know is the lifeforce. for I am Shu. Ancient Egyptian Coffin Texts. to demonstrate how the life-force enters into various life-forms: in the same way the seminal fluid travels through the shaft of the phallus. personal communication.” Mujilu can also mean “vein or artery” and can also be used as a metaphor for “penis. For the life-force (spark) comes through the male phallic shaft and enters the woman’s womb which is the incubator of life. The erect phallus is just to symbolize. Wsr as the Creator who engenders the Life Force energy into creation. We mentioned earlier that one rendering of Wsr in the Tshiluba-Bantu language is Mujilu “the sacred one.”83 We can see this association between Wsr and the phallus in the relief below. When the (right) key enters the lock. sun and water. Like the “blood of the lamb” (Yeshua). water and the sun. cited in Murdock (2009: 328).82 It is interesting to note that the “garment” that Jesus wore was also a life-healing garment. 83.also associated with the “creative life-force. it opens up a gateway for which the spirit of the person enters our world. This is why Wsr is associated with plants because plants are “channels” or “pathways” that carry the life-force: from the air. Bilolo. What‟s important for us to realize here is that Shu represents a “path” between heaven and earth.” In this very spell. Remember. I think these connections are beyond arbitrary and cannot be easily dismissed. He is also the “path” or “channel” for which this energy flows. The phallus is like a key and the vagina is the lock. I pg. the human being is a spirit that comes from the heavens and incarnates here on earth. Ramesses IX temple What we see here is that Wsr is at once the creative life-force in the universe. 116 . and this garment of mine is the air of life. iṣàn ẹjẹ. The womb and the phallus represent gateways between heaven and earth. It should also be noted that “vein” in Yorùbá is pronounced iṣàn. that in African communities of memory. 82 83 Jan Assman. the speaker invites: Come joyfully at meeting the god in me.” Coffin Text spell 80 tells us Shu is also “everlasting. This is why we see this s/s-r root in the words for air. borrowing images from nature (and yes we are nature too). This is because he also represents the life-force in the universe (àṣẹ).” the opener of the path to the mysterious source of knowledge” (Ford. Jsr/Wsr is reflected in isale “down below. or the path back to earth for those spirits who will be reincarnating. with their philosophy of “As above. We should keep in mind that both deities receive offerings on behalf of human beings “at the cross” or “altar” and both are arbitrators between the Creator and mankind. he has physical associations as well. Egyptian sr “to reach down.” imagined that a similar process had to take place upon our deaths on earth. blood). The brain is Oludumare and the only way to “communicate” with the brain is through the central nervous system: Eṣu. This is equivalent 117 . But it is safe to say that Wsr is the underworld which itself is simply the path to the spirit world. Among the babalawos of Ifa. In the body Eṣu represents the central nervous system or any of the veins. 1999: 157). The dwAt is the “underworld.” We also note among the Egyptians that there is a city in the spirit world with the same root: jsr. in African languages. means a “path” or a “channel. "to settle at the bottom of a liquid. otherwise your prayers may not be answered.” sílè "to the ground. The word for “veins” in Yorùbá is eṣan ẹjẹ (path of life. his circuit is the Duat The word Snw can also mean “divide split. vitality and life Eṣu is also known as eṣuona “Eṣu of the way. I demonstrate thoroughly how the dwAt. Eṣu opens the doors for communication and it is why in ritual one must address Eṣu first. Eṣu is often depicted with an erect phallus. So the dwAt is conceptualized as a “pathway” between heaven and earth. that are associated with Wsr. This dwAt is Wsr. to form grounds".The ancient Africans. and tree. Like Wsr of the Egyptians. Eṣu figure with erect phallus: symbol of fertility. All of the attributes that we just mentioned. Eṣu of the Yorùbá embodies these same characteristics.” In Yorùbá.” Space will not allow for a full treatment here.t "a town in the Other World" (Budge 90a). down".f f dwAt This is Osiris. What’s important here is that Wsr is the “way” like Yeshua is the way (the path) between heaven and earth.” Jsr pw Sn(w). because Eṣu represents paths and channels. so below. This is important to note because all of the “gods” in the Yorùbá pantheon are represented by human body parts (or plants). In my upcoming publication on Ogún. 1880-1920. Elegbara among the Yorùbá is the Biblical angel Gabri-El (word switch.to praying to Jesus in the Christian tradition or ending one’s prayers “in the name of Jesus” as all prayers must go through him. one of the paths of Eṣu is known as Elegbara. a path.com/akbar. We mentioned that Eṣu is also depicted as a female. metathesis)." In Amharic gebre means "servant. The protruding breasts of the above female Eṣu figure is equivalent to an erect phallus (philosophically). and that way is Eṣu (Yeshua). This term consists of two words El “God” + agbara “power”(Hebrew Gebuwr-ah “power. El-egba Legba Ne gba (spirit). Wooden kneeling figure representing the god Eshu. Nigeria." and a gibbowr "powerful man.84 84.” Ebira 'Ne Gba' “spirit”). Gabriel is also known as the messenger of God: Gaber-iy-el "the gaber of God. most powerful." Yorùbá: Fon: Ebira: Owerri Igbo: Onitsha Igbo: Elegbara.com. obi-negba "great spirit" is God) Agbara Agbala What’s interesting about this correlation is that according to SalmanSpiritual.html. All communication must go through a channel. At least I haven’t witnessed it). not God directly.” Igbo agbara “powerful oracle. This is important here because when the Arabs say Allahu Akbar which means “Allah is the Greatest." In Yorùbá we have egbere "gnome" and alagbara "a powerful man. The vitality of the woman is rendered as erect breasts (although I don’t think in real life breasts actually get erect." This g-b-r root in Hebrew lets us know that he is not only a messenger of God. 118 . In Yorùbá. It is a symbol of the life giving essence found in breast milk. El/Olu/Ala all mean GOD (proto-bantu *y-ulu). but a geber "valiant man. Retreived January 2011.” they are invoking an old African god: Eṣu. This is so because Eṣu is the one with access to all the channels of blessings (and curses). Eṣu represents the balance of nature. Eṣu safeguards the principle of freewill. following suit with the Hebrews. Absolute balance of nature.” 119 .The phrase 'Allahu Akbar' is the opening declaration of every Islamic prayer and is a slogan which was prescribed by the Holy Prophet Muhammad (upon whom be peace) to the mujahids of Islam. Eṣu is the patron of the "underworld" and their way of survival. Because all three deities are associated with water as mentioned previously. it would only make sense that they were also associated with the water course for which the liquid flows. Eṣu counterbalances aspects of our reality. However. but kept his association with the opening of prayers. He is the messenger and the message will not get to its place of destination unless there is someone to bring it.” Linguistic Connections A linguistic connection can be made with our three mentioned deities whose root is s/s-r. Eṣu is called the divine trickster that lures man's emotions creating variety which spices life. Eṣu brings out the fool in man. They in turn demytholized Eṣu. Eṣu is the gatekeeper. white and black. the guardian of the door. Eṣu is the keeper of Ase. Day and night. This cannot be a coincidence. Eṣu is an old man and a child. Eṣu brings out the symbolic child in man. abandoned the notion of separate forces doing work on behalf of the Creator. the Hebrews did not have any representations of Yeshua. Eṣu is the Divine Messenger between God and Man. A famous praise poem describes the majority of the attributes and epithets that are associated with Eṣu. Ase O! Because the god YHWH did not allow for graven images. Eṣu has a constant drive and is always ready (erect penis). Yeshua (Wsr. Eṣu has a voracious appetite. Eṣu is the Òrìṣà that offers choices and possibility. This is another correlation that Christians cannot chalk up to “coincidence. Eṣu) is still the path to God as the mediator between him and mankind. The connection with rivers is usually rooted in terms with the meaning “go” or “come. It is highly unlikely they would have adopted the phallic imagery associated with Wsr and Eṣu as they were trying to separate themselves culturally from Africans. Eṣu sits at the Crossroad. they kept the essential meaning of the deities in associating Yeshua without the graphic symbolism. construction and destruction. Eṣu's mischief serves to wake a person up and teach them a lesson. Eṣu---The means justify the end! Eṣu must always be appeased first. The Arabs. All prayers among Ifa practitioners open with an invocation to Eṣu-Elegbara first before proceeding with any aspects of the prayer (or ritual). z < d. PWS kia (+ nasal) “to go” PWS na “to come” PWN GWÌA “go”. se-gerè “to go” Mangbetu hi “road” Eastern Sudanic: Old Nubian ki. CI) “village”. sīla “road” Sumerian sila.j “come” “Bantu” gira. Bantu ji. guá “gate”. PWS CÍ “country. ĝi-na “go” PWS g i. gin.” Another variation of the term is PCS *si (< ki). Observe below. Djerma ka “to come”. the following common sound shifts should be kept in mind: s < k. kire “to come” Central Sudanic: Madi eki “to come” *G = ĝ *I = i *N = n *K = ĝ The word *gi. Masai lo “to go” Chari-Nile: Lendu ra “to go” [Maori has ara “road”}[ -an is an article] The s. Mande sīra. We can see below that our s-r/s-l root can stand for “road. kazi “road” Mande sīra. Merarit ka: ka “to come”.” In ancient and modern times. Saharan: Berti kai “to come” *K = k *A = a *K = s 120 . “motion” also *di. So the association is appropriate. sīla “road” Chari-Nile: Murle kakun “to come”. Swahili njia “road” Upper Rufiji (Johnston) gazi. gia “go” PWN GWÌA “go”. sil “road” Mangbetu hi “road”.)Nilo-Saharan: Songhai: Gao. PWS gi.” A river would simply be a “water road. Nyangiya kats “come” (imp. KI (KYI. š < t. jira “road” Mande se-gberè.(<k) could be a prefix that was dropped in most languages. ground” Bantu gi “go”. *gia means “to go” and has become a term used for “road” for which human beings walk or “go. 2009b). ROAD. gen. g ia “go”. Or what we could be looking at is a word with two roots. the rivers were the highways of travel and commerce. both meaning the same thing. When reviewing the following information. Ngala njila “road”. KAK “chase” Bantu k(a) “go” Bantu gi “to go” Bantu jidá “road” Kongo. *li (CampbellDunn. *ri. oro “to flee” PCS *la “motion” Eastern-Sudanic: Merarit la “to go”. Tama lo “to go”. JOURNEY KA “go” [R] Sumerian kaskal “road” -s -l PWS gà “place”.GO GI “go” NA “come” Sumerian ĝin. Ngala jila “road”. laba (1).] ** L = r *A = a *K = # *A = a Sumerian hara “miller” (ara 3. Kalungu. Ngala etc njila “road” We see repeated sila and jila in some form for “road. This root is found in the Isizulu word njalo “perpetuate. 370 ja “road” (209). The following may be relevant as “roads” are perceived to be “long. In the ancient Minoan 121 . E. GI “go” PWS la “earth”. with frequent reflexes in ya. Bangwelu. Cameroons-Cross River kuri*. Luwumbu. eternity”.] [Johnston (1922 : 371) has jar “road” (253). KI. [Malay has jalan “road”. Cameroon. with final n. S. Linear A JA means “gate”. DA. sika* “long”. jidá “road” Kongo. kucu. Lufula. Luchinda. leak” Sumerian íd-a “river. Swahili njia “road” Bantu jidá “road” Mande sira. Felix Chami in his work the Unity of African Ancient History: 3000 BC to 500 AD speculates that the Egyptian god Ra may have been a water deity in remote times. Mande kunu “long” (garment) Afro-Asiatic: Chad: Podokwo (3) guda. canal” Bantu gi. KUA “go”.” Variations of ru and lu in Central Africa appears in names for rivers and lakes such as Lualaba. Bantu (k)a “go” is the basis of this “prefix”]. to “go” on a course continually (used in a Zulu blessing: ume njalo “may you stand forever”).root for road is the same root for “water” and “rivers. PWS gía “to go”. la (199). Nigeria. p.ROAD LA. A road is a long path for which people. in other words. Cushitic: Galla (E) guda “many” The same -r. 2006: 69). Luangwa. di “river”. “below” PWS là “earth. k(a) “to go”. lak’ “to march” Kongo. and Lufuko (Chami. ala (204. PWN GWÌA.” It could be #gi (>#si) “go” + #ra “road”: to walk a path. ground” a- Sumerian a-ra “road” Sumerian ra “go” Sumerian ha-ra-an “road” Bantu (Johnston 1922: 535) lāā (16). 5 ) “grind” Nyanza (Johnston) 4 a-b izi. Rhodesia izi “river” PWN DU “drip. Linguistics may back him up. Lukulu. sila “road” Mangbetu oda “to proceed” Eastern Sudanic : Merarit la “to go” (elsewhere lo) Chari-Nile: Lendu ra “to go” [Bantu (k)a “go” (compare Sumerian ka “gate”) for the prefix. animals or water (even lava) travels. Wu-Nyamwezi gulu* “long”. laγa (120). The root is “go”. RA “go” KA. 205) “go” “Holoholo” lah “to flee”.” Northwest Zambezi suku. South Sierra Leone şul “long” [Alternation of l/r/d]. Lubu. donga “river”. (da) “day” (sky = rain) PWS gi “firmament” Gur *do “sky” Mande la.i) “to eat.-e “water” PWN BUDA “rain. waterplace” Bantu dé “sky. PWN DU “drip. rain”. (di) “head” PWS lim (dim) “extinguish” (of rain from the sky.. leak” subterranean water” 122 . Lotuko (na)are “water” CN Berta ro: “rain”. Tama ar “sky.ko “water pot” Swahili mto “river” Mande kulu-ntu “winding river” Mangbetu tu “pond in the forest” RAIN DA.mb “rain”.nde “clouds”. Afema a-su. tua “water”. Ngala (2) še: “drink”. “cloud ”. THU. tua “water” PWN THU. clouds” PWS tu. canal” PWS li. CS Lendu ra. to “river” Kele use “sky”. THUA “water place” Bantu to “river” Bantu tu. THUA “river. We have the following in African languages to support this information: INUNDATION LA “water”. Ngombe buse “sky” Mande sã “rain” Mangbetu tu “pool in the forest” ? Mangbetu ro “sky” Afro-Asiatic: Chad: Hausa (1) ša: “drink”. Mano de “day”. ie. cloud” Bantu tu. DIM “extinguish” Sumerian idim “spring. Gio de “day” PWS ta “sky. to drink” (action of head) Sumerian íd-a “river. DI “water” Sumerian šèĝ “rain” -ĝ *A = e [R] *G = g PWS lì(d. rain. the “top or head”) ? PWN DI “to eat”. du. Guang n-tśu’ “water”. raincloud” PWN TU “cloud”. 2006: 78). “sky” PWN BUDA “rain” Bantu bú. arra “sky” *L = r *A = a *G = g/h Sumerian ra (-g/h) “inundation” -g/h PWS tu “water”. du. Dinka uar “river”. the word Ra means “snake” and “water” (Campbell-Dunn. (da) “day” Mangbetu ro “sky” ES Afitti araŋga “rain”. Logone (2) se “drink” *T = š LI. TA “sky” PWS la. (da) “day”.writing script.da “rain” PWS la. rain”. Dinka uar “river”. Tama ar “sky.m “extinguish” Nyanza (Johnston) 4 a-b izi. I believe the ancient Egyptians realized these doublets and that is why they often “merged” deities or concepts with the same attributes.Bantu dí “to eat”. iSrw “water meadow”. arra “sky” Mande la. **R = r *A = a *I = i Sumerian ra “deluge” As we can see here. (dā) “day” (sky. di “river”. This also tells us something unique: that the word w-sr and the word n-Tr / n-tr derive from the same parent (itrw “river. Rhodesia izi “river” “Holoholo” lil “to weep”(weep rivers of tears) Mande du-mu “to eat” Mangbetu ôdi “to cook in water” Afro-Asiatic : Cushitic : dimena “cloud” (under “rain”) (Greenberg) [ The verb “to eat” was also used for drinking. da “day” Mangbetu zoro “rain”. rain) Eastern Sudanic: Afitti araŋga “rain”. dí. these are very ancient words and have morphed in very unique ways. This is what we call in linguistics a “doublet”: two forms of the same word in the same language. A good example is with the sun-god Ra: 123 . S. iSw “spittle”). as they all have been known to morph into each other in African languages. Both refer to water and rivers as discussed earlier. water course. ro “sky” Nilo-Saharan: Fur ara “rain”. What I gather is that water (and its associative forms) can be rendered with the following roots: k-r d-r t-r s-r j-r All of the renderings of the initial consonant are to be expected. DA. Lotuko (na)are “water”. [Linear A has RA2 “water”]. They would often unite forms of nature that had the same or a similarly pronounced name.] *D = d WATER LA. RA “water” PWS la. canal. stream”. Bari kare “river” Central Sudanic: Lendu ra. Dr. 85 Mangbetu ri “animal or bird”. move in a “wave pattern. We may be looking at the first attempt at a thesaurus or encyclopedia. synthesizing world phenomena in very creative ways. PCS *la “motion. Sumerian ra “go”. showing sunbeams etched into the stone as waves. There may be another hidden layer to this imagery. What appears to be driving the connections. Charles S. day” ra “hawk”85 (wn) ra “high priest in Letopolis”86 Each aspect of the image below is associated with the word ra. 1998: 68). Eastern Sudanic: Merarit la “to go” (elsewhere lo).What we are seeing above is the name ra “written” four times: Egyptian rrw “snake” r “snake” ra “sun.” It is apparent that the Ancient Egyptians knew light moves in waves when we examine how they depicted light rays on temple walls. African people are good at making connections. This example of the many forms of “ra” goes to demonstrate our initial thesis concerning the unification of themes into a singular character in African myths based upon similarity of sounds and shared roots. honourable estate”). With the exception of the human body aspect of the image above. 86 124 . Songhai k’úrò “bird” In Niger-Congo la means “old” (refers to elders. for the rendering of ra above is the way in which these objects move: Chari-Nile: Lendu ra “to go”. fame. all three of the other renderings. theoretical and symbolical (Finch. in this instance. those in high rank and gods: Yorùbá Olá “elevated status.”Science in the Nile Valley represented a complete interpenetration of the empirical. Western science did not “prove” light traveled in waves until the 17th century. when in motion. Finch in his book The Star of Deep Beginnings: The Genesis of African Science and Technology (1998) plate 20 shows an image of the aperture in the roof of the sanctuary in the chapel of Hathor temple at Denderah. including the spirit worlds from which we came.In summary. Wsr (Mujilu) and Yeshua. Eṣu. 125 . the world around us. not because he was a living person and the son of God. Yeshua became the mediator for the Hebrews. We add this correlation to an already long list of direct correspondences that cannot be dismissed on a whim. but because he was an adaptation of older deities found in Africa that were the personification of concepts built on the s-r root in African languages. Wsr and Yeshua are the personification of paths and the transport of things along those paths: in the human body. the way to the Creator is through his designated channels the ancients called Eṣu. Eṣu and Wsr. but to engage it critically and systematically to eliminate or highly reduce any chances of “coincidence. therefore. But when we line up the correspondences side-by-side. one must leave the skeptics at the point beyond which they are unable to go and proceed with one’s larger enterprise. So if we find one characteristic in Jesus that is shared with Wsr. is that they all share the same roots.t (Egypt) X Oṣun (Nigeria) X Esu (Nigeria) X X X X X X X X X X X x88 X X X X X X X X We are keeping in mind that Wsr and As. Shared Characteristics among mythological figures 87 Yeshua 1 2 3 4 5 6 7 8 9 10 11 12 13 Cross symbol Life principle King Lord (of underworld) Associated w/Water Associated w/Fish Associated w/Blood A Messenger Takes one’s burdens/offerings Fights injustice Divine Love Healer God of wealth/prosperity X X X X X X X X X X X X X Wsr (Egypt) X X X X X X x X X X X X As. I consider this a minor correlation and thus the small “x” 87 126 . one can never prove any point to the satisfaction of a skeptic. we get a totally different picture and the likelihood of “coincidence” is reduced to virtually zero.” What we have presented throughout this work is powerful evidence for an earlier shared tradition in Africa for the Jesus myth in which Christian theologians will have to confront from this point on to be considered credible. 88 If we count the fact that Eṣu is the life-force found in blood and the veins for which blood flows.t. But proving the African origins of the Hebrew faith is not the end of this project. Concerning the evidence. The aim was not to look at possible correlants flirtatiously. but not As. one is satisfied that one has carried one’s investigations to the point of satisfying oneself of the adequacy of the evidence. Once. There is a larger mission involved that will be readily apparent by the end of this presentation.t in Ancient Egypt and Oṣun and Eṣu are two sides of the same principle which have been given masculine and feminine characteristics for the sake of the myths. it doesn’t matter as they are both the same principle. linguistically. or set of principles. What’s important. There will continue to be Christians who will vehemently deny the correlations presented in this work because they will look at this discussion in isolated fragments and attempt to make an argument.CONCLUSION Our preliminary journey has allowed us to explore many correlations between the gods Yeshua. The aim of this discourse is to demonstrate the African origins of the Jewish religious tradition. My hypothesis is that there was influence in two waves. Adam and Qayin). Robert Bauval’s upcoming work Black Genesis (April 2011) seems to be a promising work. III: The Linguistic Evidence. I do not count the peacock correlation as this was after Christ was alleged to have lived and there are no Biblical references to peacocks and Yeshua to my knowledge. see J. with linguistic and conceptual verification from a third. Others stopped along the For more information on the pre-desert Sahara. Sutton “The Aquatic Civilization of Middle Africa. but solid evidences of either influence or borrowing? Above I count 23 solid correspondences in two pre-colonial African nations. Many of the Old Testament figures and adventures are influenced by events that are prehistoric that were passed on as a result of oral tradition for which they chose to vilify the gods of their ancestors(i. The second wave occurred during the time of the Greek invasion into Egypt which provided the inspiration for the four gospels of the New Testament. xv. The whole book discusses the Black Africans who lived in the pre-desert Sahara who migrated and helped found Egyptian civilization. How many coincidences are we allowed before the coincidences are no longer coincidences. Judaism and the African religions shared a common shrine in the Sahara before its desiccation. It is from here that it travelled up the Nile and spread into what is now the Sahara during the Naptian Pluvial89 period when this area was green. When the area began to dry.e. pp. I think 23 correspondences eliminate the idea that the Hebrew story is wholly original and that Jesus. Each ethnic group under examination is thousands of miles apart from each other and there is no known historical interaction with ancient Hebrews at the formation of these myths.527-546.G.. The ultimate origins for the Wsr/Yeshua/Eshu mythologies lie in predynastic times which I trace to the great lakes region of Africa.Yeshua 14 15 16 17 18 19 20 21 22 23 24 Regarded as a “path” or “channel” (the way) Symbolic of bread/wheat Priests required to eat bread (body of a god) Peacock symbolism Severely cut and murdered Is a Messiah An arbitrator Share the same or similar name Savior Judge of the living and the dead Born of a Virgin X X X X X X X X X X X Wsr (Egypt) X X X As. actually lived and had adventures in Jerusalem. 4 (1974). during two different time periods.t (Egypt) Oṣun (Nigeria) X Esu (Nigeria) X X X X X X X X X X X X X X X X X X At this point we have to be honest with ourselves. Charles Finch The Star of Deep Beginnings: The Genesis of African Science and Technology. among the Hebrews. Other attributes were attached to Wsr as the agricultural revolution came to take place in North Africa. Martin Bernal Black Athena Vol. 89 127 . the ancestors of the Jews migrated east towards the Levant.E. Murdock (2009) provides some really good commentary and evidence for the latter.” Journal of African History. in the desert west of the river there is evidence of an increase in population and of pastoral societies that built large stone megaliths and sculptures. the very north of Kayinga. sometimes using themes that also appear later in Egypt.Y. Film No. Some Bantu cultures have oral traditions which trace their origins in the Sahara at the end of the pluvial period which forced the great migrations all across north and central Africa.C.nationalgeographic. Neither group stopped until they reached a viable river. Cahier XVIII/13). At the same time. Sweden. Keita. 1. along with other aspects of the culture. notes in his article titled “Geography and Climate” that:90 Between 50.O. Kayinga is the name of the country [region] where lived our ancestors in antiquity…There they already knew how to weave the cloths they wore. Chad. Dr. people came from far off north. Dr.Nile and others back-migrated in areas such as modern-day Nigeria. Lidingo. S.” in Laman’s Kongo Cultural Collection. that existed during this time based on satellite data from outer space.000 B. archaeological evidence has been interpreted to suggest that the number of people living along the Nile fell. During this period a succession of cultures flourished on the banks of the Nile. due to the region's aridity. Fu-Kiau (2003:126) provides us with details of one account according to oral literature:91 A long time ago in antiquity. (20. There are rock paintings of people and animals. people did not exist in this Lower-Congo. domesticated cattle. if at all. As rains came in from equatorial Africa in the early Holocene. microfilm. and a large watershed. Below is a map of probably what the Sahara looked like during pre-dynastic times. developed astronomical knowledge.C.000 pp. and. and people moved into the Sahara from all directions. forge hoes and. in the north.com/geopedia/Ancient_Egypt Babutidi in “Bantu Migration and Settlement. The map illustrates possible rivers. Congo and Sudan. 1914. After the climate again grew more arid after 6000 B. likely. Between 10. they come from the north of the country.000 and 15.000 years ago the desert area west of the Nile was inhabited sparsely. 90 91 128 . genetic biologist out of Howard University.000 and 6. there is evidence for migration back into the Nile Valley. There also. the desert became less arid. made the earliest known pottery in Africa. these are people who live in the Nsundi area [the left shore of Nzadi] and others are those who live on the Simu-Kongo [the right shore of the Nzadi]. and 3) natural disasters. Kairn A. They suffered a lot for this. decades before any published data that I am aware of speaking of these aqualithic cultures in the Sahara. Only three things cause mass movements of people at once: 1) “jobs”/economic opportunities (as the case with Europeans and the “discovery” of the new world).”In the past. because we have a lot of hunger up here. In this work she posits that the Bantu migrations were thrusted with the discovery of iron and bananas. but they can never pinpoint what could have made such a large population move so rapidly across such an expanse of Africa. they separated on their way. crops and fruit trees they planted dried up. Early Times to c. two chieftaincies ruled this part of the world [region]. Unable to support the suffering they said to each other: “Let’s go to Banda-Mputu [Let’s pass through the dense forest. the unbreakable wall] and organize chieftaincies.1900 CE (2003).Labeled 129 . Klieman in her book The Pygmies Were Our Compass: Bantu and Batwa in the History of West Central Africa. One such example is Dr.knives that they used.” Kayinga is the name of the Sahara region in ancient times for some tribes in the Kongo. Some authors have made hypotheses that I have found untenable given the evidence. It is highly unlikely that bananas were such a staple food that it encouraged such a wide scale migration of people. some crossed the Nzadi [Congo river]. 2) war. Linguists always talk about a great “Bantu migration” that took place 5000 years ago. Major African Rivers and Bodies of Water .” So they agreed: “Let’s go. For many years the drought reigned. The main reason for their coming in this country [area] was the famine that hit Kayinga. When people escaped from the north of Kayinga. They still have memory of the drought that pushed them south and this story was recorded in 1914. The direction of banana domestication is east to west. pp. This means that the noun-class system is not an innovation.A. It has already been stated that the Mande languages clearly represent the earliest offshoot from the parent Niger-Congo stock—not counting Kordofanian. Mande" 92 gives a probable origin of the whole Niger-Congo language family in the Sudan (pp. Welmers "Niger-Congo. It is my contention that many of the features of Egyptian theology is in fact BaNtu philosophy. Wm. But this system is not unique to Bantu languages: it is a staple of the whole Niger-Congo family. It is my contention that it is probably the Kongo area. et al. 119-120). Mubabinge Bilolo from the Democratic Republic of Congo. which Greenberg considers parallel to all of the Niger-Congo. What’s unique about this group is the uniformity of the noun-class system. This is important because many linguists posit an ultimate origin of the Bantu languages in various locations in Nigeria or the border of modern-day Nigeria and Cameroon. a bit of judicious speculation about Mande origins and migrations may not be out of order. This is where knowing African cultures and its social structures really illuminate ancient practices. Mande" in T. a westward Mande migration may have begun well over 5000 years ago. A living example can help to shed some light on this very topic. ultimately reaching their present homeland in the grasslands and forests of West Africa. Bananas are not native to Africa and came from Asia. Linguistics in sub-Saharan Africa (Current Trends in Linguistics. E. An original Niger-Congo homeland in the general vicinity of the upper Nile valley is probably as good a hypothesis as any. long separated. in his 1971 article "Niger-Congo. This will be addressed in a future publication titled Bantu Cosmology and the Origins of Egyptian Civilization. What’s important for us here is that what is a myth among modern Christians. Dr. Somewhat later— perhaps 3500 to 4500 years ago. a linguist. E. and possibly from a new homeland around northern Dahomey [now Benin]— the ancestors of the present Northern-western Mande peoples began pushing farther west. Sebeok. This information becomes important when we start discussing common cultural features that exists between ancient Egyptian cultures and that of modern BaNtu people from Central to Southern Africa.There is not enough (or any that I’m aware of) evidence to suggest that the Bantus were constantly fighting. including the Sudan. 7). 113-140 The Hague: Mouton. eds. which the origins of the Bantu lay. but a retention of features of Niger-Congo. thus they had to search for new territory. culminating perhaps 2000 years ago in the separation of its two branches and the ultimate movement of Southern Mande peoples southeast and westward until Mano and Kpelle. became once more contiguous. This was followed by a gradual spread of the Southern-Eastern division. Welmers. By way of conclusion to this general overview of the Mande languages. A full treaty of the issues and controversies of this topic cannot be addressed here as space and time does not permit. Felix Chami (2006) has demonstrated based on archeological evidence that the early Bantus were involved in cross continental trade. Wm. Perhaps the earliest division within this group resulted in the isolation of what is now represented only by Bobo -fing. 92 130 . forming a Niger-Kordofanian macrofamily. is a living reality in the Kongo and has been a living practice among Bantu and Cushitic speakers in that area for hundreds of thousands of years. Dr. The Bantu languages are only the most successful branch of Niger-Congo. From such a homeland. One of my influences and guides in the study African linguistics is Dr. The old Capital of the empire was BuKama. Dr.t) empire. we are talking about a Hrw kingdom. For example.com/Mubabinge_Bilolo. The problem is that the King also bears the title of Mfidi Mukulu-pa BuLaba/Buloba "Supreme God on Earth. of Oyá. is a citizen of Luba or the BuKama (km. who I argue is the Cheikh Anta Diop of our time. His grandfather Bilolo was a mwipu /mwihwa (nephew) of Lusamba Tshikashi wa mwa Cibambula ne Muyenge and Nkonko Tshinkenke. As mentioned previously. in the Kasai Region of the Congo-Kinshasa. Father of Nkole (Hrw). Chewa-Kingdom in Malawi and Mosambike (Mosambique).com/2006/10/kurzbiographie-prof-mubabinge-bilolo. It has in Kongo three Kingdoms : Kingdom of Matamba-Kasayi. a princess of Bena-Musawu. kulu. Their original See.) X. philosopher. Lunda-Kingdom. himself a prince of Bena-Cibondo. princedom of Kabiye ka Mutombo wa Nkole. The Primordial MushilAnga is the ancestor and founder. 2001: 201-202). etc. Bilolo’s ethnic group is called Luba-Mushianga (Mushil-Anga) which is part of a greater Bakwa-Mushilu "People Mushilu. This is important to note as I have stressed throughout this text in that the “gods” in Africa are just ancestors. Mujilu. 93 131 . and. sister of Mfumu-Kapita Lumpungu. princedom of Kabiye. Songye-Kingdom in Kasayi-Manyema. So when we talked about the Kingdom of Nkole Matamba. Lozi-Kingdom in Zambia. near Lake Mweru or Mwelu. All Luba lines when they invoke their “totem” say Bashila / Washila / Bajil (pl. His mother was Ngalula Kanku. So if I was of the Luba royal lineage. alias Bashila-Nga. Kingdom of Matamba. Egyptologist and Bantu-ologist. Wsr-Ankh in Egyptian is Washil-Anga or Mushil-Anga “Mushil-Life” in Tshiluba. Nkole in ciLuba. Throughout this text we have been making correlations between Wsr and the anx D.blogspot. the First King. Wsr can be pronounced in Tshiluba Washil.t) is the Creator of Bashilanga (Bashil-Anga/Wsr-anx). the /h/ value of Egyptian in initial position often corresponds to /k/.” In the early period of ancient Egyptian history. there was supposed to have been an invasion of Hrw kings who came from the “south.94 A lot of these names may not mean much to most because they are not familiar with African languages and its history.Bilolo is a linguist. Nkulu.” Diop speaks heavily in many of his publications on the African concept of divine kingship. Dr.speedylook. Kingdom of Kasanji-Lwalaba and Kingdom of Kazembe-Tanganyika. Mubabinge Bilolo wa Kaluka93. What is Hrw in Egyptian is Kulu. The ancient Egyptian civilization operated in a similar fashion where each “nome” was a kingdom in the federation called Ta-Mry. Ashil. Together we get Wsr-anx. What I am saying right here is that mfumu Cishila or Cijila (Wsr. alias Luba-Luluwa or Luba-Kasayi. in the Kingdom of Luba-Matamba. Mushil. nkole) denotes ancestors. He is a prince of CiNema/Tshinema nome. a nome of CiNema. in Katanga. In Tshiluba. Many of the adventures one hears in Nigeria." a royal lineage or BaShilanga Washilanga. The other kingdoms are: Bemba-Kingdom in Zambia. Ṣàngó. Kanyoka-Kingdoms. according to the Edfu text of ancient Egypt.) or Mushila (sg. nomarch of BakwaTshidimba. for instance. Oduduwa and Oṣun are based on legends of people who actually existed. He was born in 1953 in Kabwe-Nkanka. but over time have taken on “super natural” abilities as their personalities are matched with phenomena in nature. I would say I’m a Bashila-Imhotep.html. Mushil-Anga (Mushil-a-Nganga D) is the one who gave “life” (birth) to the royal line of BashilaNga. many of the kings in the old kingdom had what was called a Hrw title (Wilkinson. This is an example in real life. near Matamba and river Luluwa. Kingdom of Nkole Matamba. and his father was Mwimpe Kaluka Mfwadi Nyunyi wa Bilolo. What’s more important here is more so “who” he is rather that “what” he does. Dynasty of Kalamba Mukenge (-a-Tunsele). ciShila and ciJila. Kingdom of Nkole Matamba.html 94 The Luba empire is a Federation of many kingdoms. The very word god (*godo. in the Kongo. “Christians” (Mashiakh. the Hebrews made ‘adm the vessel for which sin entered the world. While the Africans tried to develop the Mushil in all of us. only one man can have all that power: Yeshua. Jesus is the name of an ethnic group and a royal line in Kongo. without the slightest inkling of understanding of the people. I don’t think we have enough evidence to say yay or nay given that the personality is so old. one could not be an emperor without being a fiery warrior. keeping most of the traits associated with him. to find value in who they are. One had to be the violent type who blazed like lightning and boomed like thunder: thus a Ṣàngó. 12-20 95 132 . but they possessed those titles. and their diasporic descendents. In ancient times. We now add to this. Jesus (Wsr. how can a Christian argue that Jesus (Yeshua) among the To read more on this historical fact in Africa. While Christians boast being Christian by belief. As time progresses on. Those of us who are educated on the matter must call out. people are “Christian” by blood. 1984: 30). This practice survives in the new world among dance-hall reggae artists in Jamaica who take on such names as “Mad Lion. The ancient Hebrews’ ignorance has become the ignorance of modern Christians who have now made African people. While the Africans honored Ba’lu Baale. the Hebrews made Qayin the first murderer and cursed the occupation of farming. It is from Egypt that the Hebrews adopted the mythological figure. As time flowed on. their languages or customs. Africa’s Ogun: Old World and New. the Hebrews forbade the honoring of Baal even though they worshiped another form of him as El.). It is the same car. the god/ancestor of lightning and thunder (Oduyoye. the question now posed to the Christian is. 2nd Edition. what their ancestors have contributed to humanity and to end the self-hate for all things African (including themselves). For those who don’t know the difference. “How is the Christian God from the same imagination and origin superior to the Gods of other lands?” How can a Christian argue that Yahweh of the Hebrews is better than Yahwe among the Fon? How can Christians argue El of the Hebrews is superior to Olu of the Yorùbá? That is like saying my 1998 Chrysler Plymouth Breeze is superior to your 1998 Dodge Stratus. pp. The aim of this work is for Africans (the BaMalela) in general. the original name is dropped and the theophoric name remains. 1997. the trait that led to the rebellion of his officers. I would hope that this work inspires African-Americans to research more thoroughly the faith they proclaim. Every opportunity they could get. and African-Americans (the Bakala) in particular. Washil/Yeshua was a person who gave birth to the first king (Nkole) in the Kongo. community or kingdom. Wsr took on many different attributes in nature and his legacy was used to tell various stories in ancient Egypt. the Hebrews antiAfricanism. there is no difference. they chose to take shots at African culture. at every turn.” “Elephant Man” and “Ninja Man” for example. which trapped him into suicide. Eṣu) existed in the Kongo before the advent of Christianity. but decided to deviate from the original story to match the political atmosphere at the time.” “Super Cat. Whether that was his real name or not.” “Black Pantha.95 An example can be seen with the fourth Ọba of Ọyọ who acquired the appellation Ṣàngó because of his fiery disposition. Musungidila) existed in the Kongo before Christianity. manufactured by the same company under two different names. see Sandra T. The totem of the people is representative of the founding ancestor of a royal line or village. the Hebrews said no. While the Africans revered Ogún. Barnes (Ed. I would hope that this work would help stop the constant devaluing of Black people caused by Christian ideology. After reviewing this work. While the Africans revered Mo-dimo.names may not have actually been those names. anti-African as well. human personalities are named in pre-Western communities of memory after attributes found in nature.” I dedicated this work to one of the greatest African-American ancestors who has not been given much attention by African-American people: Dr. then we need to have a real and serious relationship with “Jesus. then give us back our Cross. George Washington Carver was a botanist.W. Let’s see what you can come up with on your own. Musungidila). MD. scientist and inventor who is famous for the over 300 inventions and products he created using the peanut. embodies what a Christian should be. This goes for the other proselytizing faiths of Judaism and Islam as well.W. and give us back our writing script. 2011) 96 133 . especially when the Egyptian version predates the Hebrew version by thousands of years? If African ideas. Liester. When you do create your own deity.” To the African it is an issue of “results” and G. Carver has the results to counter any claim of pseudo science. you will have to answer for the Africans why your version of their story is better than the ones created by their ancestors.W. What we discover is that although he was born in America. with no sense of his indigenous culture and language. We discovered in this work that the real Jesus is to be found in nature and if we are truly going to be Christians (Mashiakh. then give us back our Jesus and Eloah and Yahweh and create your own spiritual system using your own names for the Creator—using terms that can’t be found in African languages that came from the imaginations of African people. to parents who were slaves. A new era of change.” The era of body-shots to African people and their values is coming to an end. anthropological. (retrieved February 15. This might lend some credence to the notion that there is such a thing as an “ancestral genetic memory bank. As Malcolm has taught us. THE GREATER ENTERPRISE At the heart of this mission is the discovery of the true Jesus.com/~wenonah/new/g-carver. in my opinion. You will have to justify on linguistic.htm. To a modern Christian this would be considered “pagan. As we have discovered.” G. The plants then told him what he should do with them. George Washington Carver of Tuskegee University. philological and theological grounds how a Hebrew interpretation of African cultural ideas and values is somehow more accurate than the people who created them and live it on a daily basis. his approach to the study of nature was indeed very African. Carver claimed that he was able to do what he did by talking to the plants. It goes on to state:96 Article “George Washington Carver: Scientist and Mystic” written by Mitchell B. When you come into Africa to do missionary work. but hate the roots.hbci. “You can’t love the tree. He. If being Black and African is so bad. that it is a curse to be who God created us to be. is among us and the old ways of thinking and its approach to human relations can no longer sustain a world as intimately connected as the world is today. its ancestors and cultures are so bad.. please make sure that he does not possess any of the 23 attributes found above. An online publication discussing the accomplishments of G. stop circumcising your sons. on all levels. Carver stated the following as concerns to his method and how he was able to extract the information from the various plants for which many of his inventions derived.Hebrews is superior to Jesus (Ashil) among the BuKami (Egyptians). What many people may not be familiar with is more so how he was able to create the products that he created from the various plant materials found in nature. No longer will educated people take the Hebrew’s words and ideas at face-value. but from a deeply personal relationship to his God. In discussing the sequential process of speech among man (thought.” He states: Let me point out. "How do I talk to a little flower? Through it I talk to the Infinite. is smelled. (emphasis mine) Although Dr. His principal interest lay in the medicinal properties of animals." Carver also spoke with the subjects of his scientific inquiry. Anything at all that existed—even a dead leaf—interested Ewurare. he was other-worldly and heavily involved in the practice of magic.Carver's mysticism sprang from his conviction that nature held the answers to all of life's questions.” Carver's scientific discoveries originated from a rich inner life built upon a strong faith in divine powers." A fellow professor and close friend. In reporting about the sacred arts of the most intelligent king of the Edo community of memory. is tasted. every hour. He was always seen to have a flower in his buttonhole. One of his earliest revelations was about his own destiny. The infinite is not confined in the visible world. a knowing in which the entire being is engaged. and he explained that he talked with the flowers and they revealed their secrets to him. The same methodology Carver used to extract information from plants is the same method the ancient Africans of the Nile Valley employed when trying to understand themselves through nature. then words). “I love to think of nature as unlimited broadcasting stations. His spirituality came not from participation in a particular church. In the same way.’ This understanding is still prevalent in many African societies. As Carver stated. is touched. His humility was maintained by avoiding what he called "the 'I' disease." he said. and the supernatural. Amadou Hampâte Bâ (1981:170) of Mali puts this concept into perspective. every hour and every moment of our lives. he elaborates on how this process is regarded as the “materialization or externalization of the vibrations of forces. and that the only requirement for obtaining these answers was a receptive ear. through which God speaks to us every day. trees. It is not in the earthquake. It isn't the outer physical contact.” Everything you witness in nature is God’s ‘speech. birds. "I love to think of nature as unlimited broadcasting stations. small force. it isn't that.” The Ancient Egyptians called this M-dw Ntr (Tshiluba M-adu-a-Ndele) “Words of God. the wind or the fire. Carver was a Christian. through which God speaks to us every day. said that Carver's gift of speaking to flowers sprang from love. "I have risen regularly at four o'clock and have gone into the woods and talked with God. Carver said. which he later fulfilled in its entirety (Eweka 1994: 66). that at this level the terms ‘speaking’ and ‘listening’ refer to realities far more vast than those we usually attribute to them. I gather specimens and study the great lessons Nature is so eager to teach us all. Everything had something to teach the young prince.’ It is a total perception. They did not only teach him their own languages but also initiated him into the various mysteries of their own natures. humility. No. I hear God best and learn my plan. since speech is the externalization of the vibrations of forces. his methodology and his philosophy was typically African." Carver experienced awe in his encounter with the natural world. When people are still asleep. It is said: ‘The speech of Maa Ngala [the Creator] is seen. and every moment of our [lives]. every manifestation of a force in any form 134 . herbs. Eweka states: Ewuare Ogidigan was greater than even Ehengbuda. though. It is that still small voice that calls up the fairies. There he gives me my orders for the day. And what is the Infinite? It is the silent. Affairs of his father’s court had no attraction for him. and acceptance. is heard. Alone there with things I love most. sound. Glenn Clark.” This ideology expressed by Carter is ubiquitous across Africa. insects and reptiles. witchcraft. Early in his life. "All my life. The Yorùbá people call this revelation Odu Ifa “the scientific study of change and the wisdom (Odu “words”) of nature. that of the power of the word (Egyptian hkA. I argue that this is symbolic of the role of trees to the earth: that it is the “backbone”. olive oil" may be of relevance here.tw/Ddw "olive tree. In terms of metaphor. pole" As I have demonstrated in Imhotep (2009). Ashe) after he was assassinated by his brother Set (Ashby 2001.97 The olive branch is often symbolic of peace around the world. really was and its importance to the survival of ALL life on this planet. In the myth of Wsr. Jesus. Jesus was a Babalawo. Psalm 52:8) and Jesus himself. Budge 1967). They understood what a plant. staff of life.g. But I did find one entry that may have relevance to us here: xt n anx (a type of plant). Wsr is often depicted as the Dd pillar which is believed to be a spinal column. The Dd/Ddw pillar is said to be symbolic of Wsr's spine. It is from the death of the plant (found in trees and on the ground) that heals the sick body. embodiment of food. tree of life" xt "wood. The tree's root is alleged to be symbolic of God and Jesus is symbolic of the trunk for which Israel are its branches. The root DD also means “a tree. was a mashiakh/musungi/musungidi “a healer. stick. that which holds up life on earth. It appears that the olive tree was sacred to the Israelites and was symbolic of Israel (e. a wood. Ese. One could speculate that this may be a secondary meaning of Jesus hanging from a tree. Jeremiah 11:16. a Hogon. woodland." Ddw "olives.” Jesus was a healer who used the methods of Africa. George Washington Carver was speaking to the real “Jesus” found in nature: the plants. That is why everything in the universe speaks: everything is speech that has taken on body and shape. Since we know Wsr is connected to plants/trees and the anx. Coptic hako.html (retrieved February 15. mast.t (Isis. Hosea 14:6. 2011) 135 . the word bAk (maringa arabica) "tree" is the same word for "spine" bAkw. the Dd is found inside a tree by Is. a Sangoma. timber. This meaning would correlate nicely with the meaning of “Jesus” in Africa: a plant. It is interesting to note that in the Ancient Egyptian language. Yeshua/Yehoshua here represents medicine that derives from plants. I have speculated to myself that there may be a conceptual link between the anx and trees. olive trees. tree. shaman”). magician. an Nganga par excellence.whatever is to be regarded as its speech. the root of the word anx is kA. Duala (Cameroon) –hango “words of power.prophecyfulfillment.” The Egyptian word Dd. When I searched various dictionaries I found the following entries: 97 See full discussion here:. a Jesus.com/en005a. Eshe. We know that Wsr (Jesus) represented all plant life which speaks to why Wsr was so important to the ancient world. I couldn’t find any direct dictionary entries for trees that possessed the anx emblem. above all other things. or tree. trees are human beings. Both figures among the Zulu and Yorùbá.com/about/biography-05/ 98 136 . This would be the stem of the plant. According to Robert Farris Thompson in Flash of the Spirit. the trunk of the tree. then it is closer to the proto-Western Sudanic version of the word *gil “tree. it can be deduced that Eṣu is the personification of the anx: just like Egyptian Wsr. (GHIM) “be alive” Mangbetu ga “the biggest tree” Mangbetu ki “wood for heating”. respectively. See Credo Mutwa’s article “Mysterious Africa the Mystery of the Cross” retrieved here:. we can equate the plants with Eṣu as well since he represents the “channels” or “paths” for which nutrients travel. A few African stories may give credence to the notion that the anx is related to trees (which I say via the human being. and like Eshu. aki “to break” Mangbetu gi “kind of tree” (protection) Afro-Asiatic: Chad: Njei kadi “tree”: Egyptian xt “tree” The A sound could have been either an /l/ or an /a/ sound. vein” PWN GHI. (Thompson. Thompson observes that: …both Eshu and Osanyin share the attribute of one-leggedness. We’ve already connected the anx with the thorax bone which appears to have one “leg. gíl. According to Zulu sangoma Credo Mutwa. among the Zulu. and Eṣu is the lifeforce in all things. a walking tree). “the one legged one. the symbol of the anx is associated with a story that states the Supreme Being’s son lost his leg in a battle with a dragon (some say a crocodile) and they call the sign Mlenze-munye (Mlente-mmunye among the Swazis) which means.” Since Osanyin is connected with plants. If an l in this case. (git) “tree”. are represented as one legged and they both lost one of their legs as a result of a fight with a divine force.”98 This is interesting because there is a comparable story among the Yorùbá with the god Oṣonyin who Neimark (1993: 159) informs us is “one-legged” like the trees and plants he represents. PWS ti “tree” PWS gi “root.” If we equate the anx with trees. But the connection doesn’t stop there. In the myths it is said that he lost his leg and acquired other physical infirmities while in a struggle for power with Orunmila. Eṣu is also depicted as having one leg.” The word anx means “life” and also stands for “human being. 1984: 43) With Eṣu being the crossroads and with him having one leg.Tree in Egyptian language kAA (part of a tree) kAA (a type of plant) kAkA (a tree) These entries stem from old African roots for tree: PWS gí. Osanyin was once a prince. the relationship to “life” and “human beings” is still maintained because in Africa. “What is this?” I said. I praise it. grandfather. The ancestors are the real school of the living. This world is where one comes to carry out specific projects. t>d). which was a fig tree that he had planted as a young man and which was now tall and producing a lot of beautiful fruit.” Malidoma Somé demonstrates the roots of identity in African communities of memory. therefore. They are the keepers of the very wisdom the people need to live by. Mangbetu ti “trunk. Wsr and Yeshua are all associated with royalty: either as Kings or Princes (Jesus being king of kings AND the prince of peace). Bantu tí “tree”. Mutwa proceeds to retell the following account of what happened when his grandfather asked him to follow him after giving him his first lesson: I [Credo] followed him [grandfather]. Osanyin. I have seen you touching the trunk of this tree. a spirit who has taken on a body. Grandfather then proceeds to initiate him. My grandfather said. a gap-toothed. every person is an incarnation. “The family is like the forest: if you are outside. “Have you seen me standing next to this tree on certain days?” he demanded. I was worshipping the tree. every tree has its own position. He said. “Now what did you think I was doing. “Listen. rather harshly. Do you understand?” I said. it is dense. we never used to call trees ‘trees’ but rather ‘growing people’. I thank it—and see the fat figs that it produces for us because I talk to the tree and I believe that it is a person.” My grandfather laughed. A Kongo proverb states Nsi mfinda “Nations are forests. grandfather.” (Mutwa. by way of the Dagara. At this point in the story Mutwa is asking a lot of questions concerning the nature of God and the contrast between indigenous spirituality and that of his Christian faith. rivers and still water. mountains. I was talking to the tree. grandfather. you little Christian rubbish? Did you think that I was worshipping the tree? Did you think that I thought the tree was my God?” I said. and I often share any good news that I happen to have with the tree. The life energy of ancestors who have not yet been reborn is expressed in the life of nature in trees. First. Grandfathers and grandmothers. I said.” My grandfather struck me across the face and said. All four deities share the same s/s-r root and are associated with plants. I was sharing my snuff with the tree. “Listen you little dog. Bantu di. “Yes.” “What have you seen me doing here?” he asked again. this is a person. A birth is therefore the arrival of someone. “Yes. if you are inside. My grandfather said.” An Akan proverb states. who has important tasks to do here. “Grandfather. This is a person. Do you understand me? In old Africa. that is. when he states: For the Dagara.So we observe here as well that Eṣu. I sing to it. So our true nature is spiritual. trees are considered people in African tradition and can be communicated to like human beings: PWS tí “tree” (“tree of life”). into the wisdom of his ancestors. “tribe”(tribes being thought of as “trees”. in the land of the ancient Zulus. “No. “Grandfather. origin”. 2003: 15) From this account we understand several things. and at one time I saw you taking snuff out of your snuff horn with your snuff spoon and pouring it at the foot of this tree. usually an ancestor that somebody already knows. Credo Mutwa in his book Zulu Shaman recalls an initiatory event as a child that involves him learning the wisdom of his ancestors as told to him by his Grandfather. He came to his favorite tree. in my time when I was a young man.” Another blow flew across my face and snapped my head back. His grandfather is informing him of the Zulu belief in a symbolic “great lake” of hidden knowledge in the spirit world which is accessible to human beings. cruel laugh. this is not a tree. are as close to an 137 . this is a tree. expression of ancestral energy and wisdom as the tribe can get (Somé 1994: 20). (emphasis mine) I am currently editing an anthological work by Dr. Baruti Katembo titled Scattered Assets: How African-Americans and Other 21st Centery Resources Can Aid African Empowerment (forthcoming summer 2011). In this work Katembo (along with others) has an article titled “Mathematics, Circularity and the Msonge” that discusses the viability of circular homes (Kiswahili msonge) in Africa and the relevance of circular design in space optimization and in engineering efficiency as a link to socio-environmental perception and understanding, with the emergence of Convergent Evolution Theory. In this discussion, Katembo notes that architecture reflects the culture and distinct civilization elements of a given society. With that said, he noted that the famous “hut” homes in Africa (the thatched roof dwellings) were imitative of trees with hanging branches—thus an extension of the natural space. We are beginning to see just how important trees are in preWestern communities of memory. We’ve already noted previously that the forest is considered the spirit realm (the place of one’s ancestors), that trees are people themselves, and now that the homes were imitative of trees which would make a village a forest (Bantu di<ti “tribe”; tribes being thought of as “trees”). Secondly, Mutwa acknowledges the “worshiping” of the tree. The word worship simply means “to honor, love, venerate, revere and respect.” Christians have attempted to make the word worship a negative term when not directed at what they deem the ONLY thing worth loving, honoring, revering and respecting. If they “worshipped” nature in the way Africans did in ancient times, maybe we wouldn’t be having a water crisis at this time, or polluted air, or nuclear waste in the soil, or have genetically engineered food, or many of the diseases we have as a direct result of the loss of plant life and damage to the environment due to the greed associated with capitalism. Tree of Life (medicine) A lesser known òrìṣà among the Yorùbá is the deity Osonyin who further illuminates the relationship between our s/j root shared among the deities compared in this text. The root of the name Osonyin is the word oso which stands for a “seer, magician, herbalist;” in other words, a priest or healer. Osonyin is the òrìṣà of plants and the inherent àṣé and magic they contain. Stated plainly, this deity personifies the life power in vegetation and medicine. This òrìṣà is obligated to the aje or witches. Like Wsr of the Ancient Egyptian tradition, Osonyin is also associated with a tree (the Araba tree). He is the owner of the forest as well; which in Africa is the universal home of the ancestors and God. The babalawo’s staff, or Osun (not to be confused with the òrìṣà of the same name), represents his pact with Osonyin to utilize the properties of herbs for magical purposes (Neimark, 1993: 160). In the Ifa tradition, medicine is known as Ewe, which sounds awfully close to the spirit of nature called Ewa in the box office film Avatar. In discussing the nature of Ewe among Ifa practitioners in the new world, Oduwafon in his article “Ewe Orisha: A Treatise on the Role of Plants in Yorùbá Religion” notes the following that I think is relevant to this discussion: “No hay Santo sin Ewe”, there is no Orisha (God, spirit) without plants. Trees and plants are endowed with a soul, intelligence, and will, as is all that is born, grows, and lives under the sun. “Wasn’t Jesus born in the monte on a heap of herbs and in order to go heaven to become god didn’t he die in the monte, Mount Calvary? Always walking and putting himself in the mountain. He was a ‘yerbero’”. Each plant has a virtue of an Orisha, a supernatural force. The medicines are living in the monte. (unknown date, p.4) 138 Monte is the “mountains” (the wilderness, the edge of the forests). What’s important here is that he notes that trees and plants have souls and intelligence. This is a belief that has been brought to the new world with many scientists having successes following this path of logic. A Luba proverb states, “Intelligence is a flower that grows in a neighbor’s garden.” As Malidoma Somé observes, concerning the level of consciousness on this planet: Plat’s cave. (Somé, 1999: 50) The ideas brought forth in Judaism and Christianity are incompatible with the world-view of African people in regards to nature. For Africans, human beings are a part of nature and we must learn to live in harmony with it. We are not that intelligent to dominate it. The Abrahamic perspective is vastly different. In this tradition, human beings are here to dominate, not learn from nature. If only Christians understood our relationship to nature and the real Jesus, then maybe many of the pitfalls of our historical experience could have been prevented. As Somé notes, “Nature is the textbook for those who care to study it and the storehouse of remedies for human ills (Somé, 1999: 38). If Christians understood that nature is a textbook of valuable information, and that we humans are a part of nature, then maybe they would have a different approach to human relations as they would see others, who do not necessarily believe the same things as they do, as sources of wisdom. In other words, there wouldn’t be any perceived hierarchy of ethnic groups as all human beings are seen as valuable resources for which we can mutually enrich each other’s lives. If we can learn a lot from a dummy, we can learn even more from a tree. Why we need “Jesus” in our lives. Most reading this would think that I am anti-Jesus when the opposite is true. I am PRO JESUS. I am a huge advocate for the promotion and elevation of the real Jesus: the plants. I understand fully the significance of “Jesus” in our lives and how he is being “crucified” for the “sins” of humanity. What I have a problem with is human beings not seeking to understand “Jesus” and why he’s so important to our lives. It is my contention that this is the case precisely because we abandoned the traditions of our ancestors. We felt we were more intelligent than they were and we are paying for our “sins” big time. Oduwafon in his article on Ewe Orisha “plants” underscores just how important plants are to the earth and human existence. What follows is a summary of what he presents, with slight modifications from me. In this discourse he is speaking of Osain, which is the Cuban pronunciation of Yorùbá Osonyin/Osanyin “the spirit of plants and herbs,” and their significance in Yorùbá medicine. Within this discussion he highlights some scientific aspects of plants, for which if one examines carefully the language he uses to describe the role of plants in our environment, one can detect a similar explanation of the role of Jesus in the Biblical myth. All of the attributes one could think of in terms of plant-life, one could easily substitute for the epithets of Wsr, Eṣu, Yeshua and Osonyin. It is in the scientific explanation of plants that we truly get a sense for where the true origins of the Jesus myth can be found. 139 Oduwafon goes on to note the following:99 1). Yeshua as the captor of energy from the sun. The major source of energy to the planet earth comes from the sun. A small amount comes from the interior engine of the earth which results in volcanic eruptions and earthquakes; the majority of the energy manifesting itself on earth comes from the star called our sun. This energy reaches the earth and some is absorbed by the atmosphere warming it and creating the winds (Oya) and thunder and lightning (Shango). Some of it is absorbed by the land and the oceans warming the planet and creating our climate and ocean currents. The plants through the process known as photosynthesis capture the remainder. It is this process that is responsible for life and the progress of civilization itself. The energy which is necessary for almost all life processes (there are certain living organisms which live deep in the dark regions of the oceans which get their energy from a process other than photosynthesis called chemosynthesis), comes from plant photosynthesis. All human and animal food comes from plant photosynthesis. The green leaves (chlorophyll) capture the sun’s energy (light). This energy is then stored in the Wsr first as glucose, a sugar, and then the glucose can be transformed into other organic molecules (fats, starches, other sugars) which then can be consumed by living creatures. Thus plants are basis of all food. The energy which animals ultimately use in performing their daily tasks comes to them via Jesus from the sun. Photosynthesis Equation 6CO2 (carbon dioxide) + 6H2O (water) + photon (light) C6H12O6 (glucose, sugar) + 6O2(oxygen) All fuels which man uses to heat his homes or drive and run his machines come from photosynthesis. The glucose created above, contains the stored energy that can be changed to a multiplex of hydrocarbons (compounds containing HxCy) and carbohydrates (compounds containing carbon, hydrogen, and oxygen) yielding wood, coal, oil, and gasoline. Fossil fuels: oil, coal, natural gas are nothing but decayed plants, 100 heated and pressurized for millions of years. The energy which is contained in them is the energy trapped from the sun by green plants in photosynthesis. So that all the energy burning in our cars from gasoline is energy which was created by the sun and captured from sunlight by green plants. So all of these physical, chemical processes take place under the aegis of Osanyin. 2) Osain as the giver of oxygen (life). The photosynthesis equation represents not only the trapper of the energy from the sun, but also the maintainer of oxygen in the world.101 The union of carbon dioxide and water yields glucose, the storer of the energy, and oxygen. The oxygen which is needed by all life forms is continually being created by the green trees and vegetation in the world. Twenty percent of the oxygen produced by the world is produced by the Amazon rainforest. When one enters the Igbodu, the Yorùbá place for ceremonies, a place filled with fresh green ewe, plants that will be used in these ceremonies, one feels invigorated, I have modified his notes a bit, substituting the words Osain and Plants on occasion with the names of either Yeshua, Wsr or Esu for effect. The hope is that one begins to look at the meaning behind names in myths so that one learns how to substitute the “meaning” in place of the “name” of the character to get a better understanding of what is going on in the narrative. 100 Crusified Yeshua (Wsr). 101 The breath of life 99 140 refreshed, inspired, strong, a feeling that all wrongs will be righted, that bad will go away and good will come. This is partly because the plants are creating fresh oxygen, and dispelling it into the air giving life and strength to the environment. Looking at the Osain photosynthesis equation. On the right side there is O2 and C6H12O6. O2 is oxygen, air which is an attribute of Oya. Oya is the goddess of the winds and air and since she is the goddess of air, she is the goddess of life,102 for all living things depend upon air to breath to have life. As she is the goddess of life she is likewise the goddess of death, because when one stops breathing, one dies. She is the owner of the cemetery, the gates to and the entrance to the cemetery. She presides over funerals and is the only Orisha that can dwell with the eggun, the spirits of the dead. Oya is also called Yansa, Iya mesan which means the mother of nine because Oya gave birth to nine children (Odu OsaMeji). The last of these children was Eggungun, who represents the spirits of the dead. So Oya has a close connection with death. When people are dying too much and too often, Oya is invoked, prayed to and played to, in order to stop the dying. Oya controls death. Oya is represented by O2. As air comes from the plants, so Oya is connected to Osain. Oya in a negative light is the hurricane and the tornado, air gone wild. Oya in Yorùbá means it tears, it rips. Osa, (Yorùbá) means he/she/it runs. C6H12O6 is glucose and in it is stored, as chemical potential energy, the energy that was captured from light. In glucose is the energy that can be put to use for cellular functions or to burn gasoline. Shango is the representation of physical energy. He is associated with fire, his color is red, he is one of the gods of war, and he represents male sexuality. He is the bravest warrior of the pantheon. Wherever he goes there is war, arguments, and discussions. Energy! Energy! The energy contained in C6H12O6 is a representation of Shango. It should be noted that in the second apataki for the disfigurement of Osain (Cuba) which talked of origin of Osain, there was Oya, Shango and Osain. 3). Osain as the cleaner of our air and atmosphere. The photosynthesis equation generates not only oxygen, but also depletes the air of carbon dioxide. Carbon dioxide, a dangerous gas which if inhaled too much results in suffocation, is one of the major byproducts of burning and respiration. It is also the major greenhouse gas associate with global warming and climate change. Plants are called the ‘Lungs of our Planet” because they take in carbon dioxide and exude oxygen.103 Since the beginning of the Industrial Revolution (~early 1800’s) the concentration of carbon dioxide in our atmosphere has increased 30% and the average surface temperature of the earth has increased by one degree. Carbon dioxide is a greenhouse gas which traps the heat radiated by the earth and reradiates it back to the planet causing the temperature to rise. Plants take about 12% of the CO 2 and convert it O2 as such they are cleaning the atmosphere of unwanted pollution.104 Prior to the industrial revolution, the rainforests and plants were enough to offset the carbon dioxide produced by natural processes of animal and plant respiration. However with the recent onset of anthropogenic addition of CO2, and the destruction of the forests and rainforest regions throughout the world, the situation is looking bleak, with the concentrations of CO2 steadily increasing. Green plants remove the pollutant carbon dioxide from the air, 105 and give oxygen. 4). Plants are the natural food for mankind. In today’s modern era, humans are hooked on fast foods, meats, and all other garbage. Eating plants, vegetables, fruits106 are the best source of nutrition. Eating plants, vegetables, cooked and Yahweh, Yahwe, Yeve, Ya, Jah, etc. Take in our sins, our burdens and gives us new life (oxygen, breath, air): 104 Washes our sins with the blood of Jesus (chlorophyll). 105 The sins of the world 102 103 141 uncooked is the best choice; Salads, fruits, and nuts. Digestion is easier (meats, especially beef is very hard to digest), and diseases such cancer and diabetes are diminished, disappear and longevity and strength are achieved. 5) The medicinal value of Osanyin. It estimated that the number of plant species on earth is 400,000. Currently, 121 prescription drugs sold worldwide come from plant-derived sources. While 25% of Western pharmaceuticals are derived from rainforest ingredients, less that 1% of these tropical trees and plants have been investigated by scientists. Aspirin, acetylsalicylic acid, is derived from the bark of the willow tree, opium from opium poppies led to the creation of morphines and codeine, and quinine from the bark of the South American cinchona tree. 3000 plants have been identified as cancer fight[ing] drugs. Twenty five percent of the active ingredients found in cancer fight[ing] drugs come from Wsr. More than 100 pharmaceutical companies are engaged in plant research to find new cures for infections and diseases. Most of this research is carried out by examining the practices of traditional shamans in the forest areas. When we understand Wsr, Yeshua, Eṣu from an African-Centered point-of-view, we receive a more rich interpretation and a greater understanding of the relationship of him to us (the reader of the narrative). All African myths (that I’m aware of) are not stories about external beings having external experiences, but internal realities that we must confront in order to make our experience on earth more enjoyable and harmonious. What Christianity has done is made Yeshua (Osiris) an external being, thus one does not see themselves in the narrative of Christ. Most Christians are satisfied with being “Christ like” instead of being full-fledged “Christs” (Mashiakh, Musungidi). This undervaluing of the Self has caused many human beings to not reach their greatest potential. Some people, as a result of their religious views, honestly think that using one’s brain to the highest capacity is a sin. It is this latter view that has caused many persons to not seek clarity on the faiths they practice by critically examining it using the best tools of analysis the various fields of science currently has to offer. This has been done at the peril of good human relations and truth. A pastoralist society is not about developing great thinkers, philosophers or scientists: they are about making people sheep. The whole mode of production is about domestication and it speaks to the ideas expressed in the Biblical literature of having “dominion” over nature and having a hierarchy of certain ethnic groups over others. When you have a social framework organized in this manner, rooted in this paradigm, then slavery is the only logical outcome for such a people. We continue to be perpetual slaves to Western religions because we only see Africa as a reference and not as a resource. We don’t ask Africa fundamental questions of life because we think that Africans are primitive children who need the help and guidance of Western parents. Africans are the world’s 106 The body of Christ, Wsr… 142 elders (*godo, *gudu, bukulu, hrw, ka, nkaka) and we have a lot to teach the world, primarily because we learn from nature (our elders) who teach us about life. It is African know-how that has helped humanity survive for over 2.4 million years of existence (160, 000 of that in homo sapiensapien form). It has only been in the last 2000 years that the threat of the end of human existence has become a reality. It is not a coincidence that this is around the time that Christianity (Judaism/Islam) started to dominate the human imagination: not by way of love and compassion, but by way of the sword (and musket gun). It is time that we build a new ethics, but one that is healthily related to traditional ethics. It is the ancient systems of human organization that can show us what makes long-term human survival possible. At the heart of the matter is, Christianity at its current level of consciousness, does not and is incapable of fostering good human relationships. It is a value system that encourages bad character in people by insisting on value-judging others who do not adhere to the same world-view as the Christian (or Muslim). The Christian identity is built on “othering” people: by being anti-others. A culture that is built on being anti-others is a culture whose foundation is built on sand; not in the rich fertile soils where people can grow and develop into wise sages who know how to add life to life. African cultures are community-centered. Africans from an early stage of human development realized that our best chance of survival is by working together: thus the community was born. This community not only includes the living, but the yet to be born and the ancestors who have left. In developing strategies in which to keep the community thriving, in balance and secure, they created social systems, rituals and taboos to ensure that every citizen has the opportunity to experience the most exciting and fulfilled life possible. In creating rules for communal living, they in turn created the first civilizations, as the primary objective is the complete realization of the person which requires those within the community to act civil towards each other and those outside of their immediate sphere of influence. Part of good character development in Africa is in acknowledging and respecting one’s Elders and Ancestors. It is time Christianity develops good character: respect your elders (in visionary leadership through spirituality and in nature). The Spiritual Grand Strategy I would like to borrow a few concepts from Kajangu (2006) in proposal of a spiritual grand strategy that would allow us to mobilize all the powers of our being to wage war on the syndrome of “foolishness” and fertilize our seeds of greatness: the aim of African wisdom centers. The first concept is the powers of being. African wisdom traditions inform us that a human being is what his senses make him to be: What we were fed at birth What we feed ourselves What we see What we learn/hear What we believe in What we do What is around us (known and unknown forces) What our parents poured into us before birth: our biogenetic package (Fu-Kiau, 2003: 79) 143 These traditions hold that human beings have both open senses and closed senses. The open senses consist of the following body parts; two eyes, two ears, the nostrils, the anus, the sex organ and the mouth. These are not considered organs for the senses, but the senses themselves. The closed senses consist of the heart, the occipital region of the head and the fontanel. These parts of the body are considered openings that were closed when human beings were sent into exile on earth. Initiation aims at “reopening” these closed senses so that we may be able to “see” the hidden layers of our reality. It is the closure of these senses (our hearts and minds) that prevents human beings from participating in the business of the stars (as we are children of the earth and children of the stars). The senses are what construct all of the experiences that human beings go through here on earth. The powers of being are channels of power and the ultimate builder of the self (our double and personality). The open senses are instruments of life that empower us with tremendous energy. They possess divinatory abilities that can help us make life more beautiful on earth. We are able to cross the rivers of life by utilizing the bodies fuel refinery, which itself is powered by the plants we eat. We as humans experience two types of hunger: the greater hunger and the lesser hunger. The lesser hunger has to deal with the needs and wants of life on earth. This can be satiated by material things. The greater hunger, however, can only be satisfied by looking for wisdom. It is this greater hunger that has inspired the creation of the African life principles institutes in order to maximize one’s ability at gaining wisdom. Unfortunately, these wisdom institutions have become religions in the Western imagination and satisfying the greater hunger is no longer the goal of the modern gathering of the minds. Foolishness exploits our lesser hunger and makes us violate the boundaries that define the appetites on earth. It is this violation that forces Mother Nature (mdw nTr) to retaliate and defend herself against the self-proclaimed “dominion owners” of the planet. African wisdom traditions perceive the earth as a hidden sachet (futu) of medicines, tied up by the Creator, which can help us stem the tides of adversity. Earth is the place where we reap wisdom from the ground. The closed senses mentioned previously consist of the occipital, the fontanel and the heart. The two formers are considered as the spiritual eyes and the heart is believed to be the “internal voice” of heaven. It is the spiritual eyes that allow for us to perceive the hidden layers of reality; and it is the heart that facilitates the dialogue with Heaven-dwellers. Wisdom seekers use the powers of being to explore life beyond the limits of the material universe. Wisdom seekers use the spiritual eyes to analyze and organize the experiences that are manufactured by the open senses, but also to forge these experiences into a holistic outlook that makes it possible to know life D. It is the spiritual eyes, working in concert with the open senses that allow us to see the true spirit of Yeshua in the environment in which we live. Wsr is the embodiment and union of the life cycle and its elements: water, air, fire (the sun/son) and earth. By examining mythology from the standpoint of African wisdom traditions, we can learn how to utilize the “savior” (Wsr, Yeshua, plants/trees) and become saviors ourselves (nganga, bokoles, Nkole, mashiakh, musungidila, medicine men/healers). We accomplish these lofty goals by building our own bags of wisdom. In the search of wisdom, wisdom seekers are guided to come up with their own answers. This is done so as not to be dependent and enslaved by books or spiritual “gurus” so that one becomes their own ntr. The bag of wisdom is a spiritual reservoir which wisdom seekers build with symbols, which are the elements of creation’s alphabet (mdw nTr), that enable them to know life. These are collected on the superhighway of wisdom which is a virtual network of wisdom centers around the world. It is a 144 network that makes it possible to establish a dialogue of mutual enrichment among wisdom traditions. This is what has been lost due to the Abrahamic religions. Wisdom traditions in antiquity believed that no single person is the mother of wisdom; that it takes the sweat and tears of countless sages working together over thousands of years to build a wisdom tradition. It is the belief of African wisdom traditions that a wisdom center cannot flourish alone for it needs to engage in dialogues with other wisdom traditions. It is for this reason that ancient wisdom traditions established a superhighway of wisdom which is still open today.107 A Bairu proverb poignantly captures the sentiments of African centers of wisdom: “The child who has never left home says: My mother is the best cook!” The Abrahamic religions are claiming to be the “best cooks” and have never, themselves, left home. They have not travelled and sat at the feet of masters (nganga) around the world and collected items for their bags of wisdom. They never learned how to tie and untie knots of their ancestral, biogenetic ropes: their signs and symbols which empower us to gain a competitive edge against other animals that are better equipped to live on earth. We must all use the sacred arts to develop our powers of being. They serve as a polestar for travelling on the superhighway of wisdom and for building our bags of wisdom (our collection of symbols and tools for stemming the tides of adversity in the sea of life). We ultimately study nature to learn how to solve problems and to develop a towering imagination to find inspiration for unveiling life’s mysteries and to make life more beautiful on earth. We use the sacred arts (myths, proverbs, divination, science, visual arts and song) to help us expand the diameter of human capacity and to discover new more satisfying dimensions for being human. It is these ancestral practices that inspired the myths of Wsr/Yeshua/Eṣu. When I note that the Jesus myth has its origins in the Kongo, I am not only speaking of a particular geo-political location in Africa, but a conceptual realization found all over the earth: in plants. Remember from our earlier discussion on the anx: Anga > BuAnga/Manga Medicine, remedy = the protected and cared for. Present in: Ba-Kw.Anga, KoOnga, Kongo, KaAnga The word Kongo derives from the root anga which refers to “medicine, remedies” (that which cares and protects life). Medicine and remedies are found in plants/trees. The Kongo refers to the forest, the place of initiations (of physical and spiritual growth and education). It is also a word for God: Kongo = God; Bundu dia kongo = together of god; Muana Kongo = son of god; Bena Kongo = from God. The forest is also considered the ancestral realm and the home of spirits of those yet to be born awaiting their carnations on earth. The plants are our oldest living ancestors on earth. For they produce the oxygen for all living things to breath, and without oxygen, there is no life on earth. So we, as African/African-American people, respect (worship) life by respecting our environment and the people that live in it. The idea of viewing nature as family is not only an African thing. I do not want to act Christian in this sense and try to “monopolize” ideas shared across the globe. Dr. Vandana Shiva in her article “Local Solutions to a Global Problem” underscores the need for a new global ethics, rooted in traditional paradigms that will enable us to return to a sustainable way of living that doesn’t 107 See Asar Imhotep’s article “The African Superhighway of Wisdom” (July 2009). 145 threaten the very life principle itself. When asked, “On which levels should change take place,” she responded: We need to respect nature and treat nature as our family, the way we are taught in societies like India. We need to realize that we live in the earth family. Humans are not a privileged species. We are just one more species on the planet. If we end the colonization of nature, we end the ecological crisis. (Joubert and Alfred, 2007: 280). Although I am of recent African (Dya Malela) descent and I adhere to a synthesis of world-views that are shared by many cultures on the continent of Africa, I only use the term African World-view for political reasons as Africans prior to colonization did not have an “African” world-view: there was no Africa. If I were to properly label the kind of thought process which I politically identify as “African,” it would be called indigenous. The aim of this work is to inspire the reader to think more indigenously. Indigenousness is rooted in two things: earth-centeredness and human-centeredness. The Abrahamic religions do not allow us to have a meaningful human-centered public discourse because its orientation is on following the rules of the book: a book that doesn’t allow for updating and adjusting when old paradigms no longer create harmonies in the social and natural spheres of our lives. The aim here is to develop and expand our indigenous way of thinking so we can better use our “Indigenuity” to better (and more sustainably) solve our problems. Indigenuity is a term I coined that combines the concepts of indigenousness and ingenuity which implies using our indigenous (human-centered and earth-centered) world-views, in combination with our creativeingenuity (spiritual resourcefulness/skills) to 1) discover new more satisfying dimensions for being human, 2) unveil life’s mysteries, 3) know life to the highest capacity and 4) make life more beautiful on earth. The ultimate origins of Jesus is to be found in the scientific understanding of plants and their relationship to man. Only when we are able to open our closed senses and learn to see what the ancestors have left us through myth, can we discover the true Jesus. The fate of humanity rests on us accepting “Jesus” into our lives (eating more vegetables, using natural remedies when possible, creating and using eco-friendly products and services): not in the religious sense as proposed by modern Christians, but in a spiritual and ecological sense as proposed by the Ancient Africans who created the Jesus myth in the first place. I end by recommending two books to read that I feel will broaden the reader’s perspective on the underlying issues present in this book. The first is Rooted in the Earth: Reclaiming the African-American Environmental Heritage. Dianne D. Glave. (2010), Lawrence Hill Books; and the second book is titled Beyond You and Me: Inspirations and Wisdom for Building Community. Kosha Anja Joubert and Robin Alfred (Eds.). (2007), Permanent Publications. Ancestrally, Asar Imhotep (ciShila/Mujilu/Washil ciTapa; Asalu Adabo) – Sun of the Soil February 27, 2011 146 Dr. Mubabinge Bilolo D (Egyptologist, Philosopher, Linguist) and Asar Imhotep outside of the Houston Center Mall in Houston, TX: Dec. 31, 2010. 147 Wallis. (1858). John D. (1913). ________________. B. Smithsonian Institute Budge. Johannes. 2nd Edition. and Kalamba. ________________. (2006). Volume 11. John Benjamins Publishing Company. Bimwenyi-Kweshi. Felix A. The Egyptian Book of the Dead: The Papyrus of Ani. (1997). Wm. Ayele. Cruzian Mystic Books. Eisenbrauns Allen. (2003). ________________. Dictionary of Symbolism. E. Bengston. John Benjamins Publishing Co. (2006). Renaissance of the Negro-African Theology: Essays in Honor of Professor. St. Lagos 148 . (2009b). (2010). Bilolo. Society of Biblical Literature. G. Munich.” Unpublished paper: The University of the West Indies. Dictionary of Yorùbá Language. James P. Funso. Penny Farthing Press. GJK. Dover Publications. Nsapo. ________________. Who Were the Minoans: An African Answer. Academy of African Thought.). Penny Farthing Press. Tucker. Middle Egyptian: An Introduction into the Language and Culture of Hieroglyphs. The Unity of African Ancient History: 3000 BC to 500 AD. An African Writing System: Its History and Principles. Eds (2009). E&D Limted. (2008) In Hot Pursuit of Language in Prehistory: Essays in the four fields of anthropology. (1996). An Introduction Into African Languages. Grammar and Dictionary of the Yorùbá Language. (1967). Albright. (2009). Bowen. Trinidad and Tobago.). Kinshasa Botterweck. Ethiopic. Resurrecting Osiris: The Secret Mysteries of Rebirth and Enlightment. Church Missionary Society Bookshop. Sumerian Grammar. Biedermann. (1997).Theological dictionary of the Old Testament. T. (2004). Freising. Campbell-Dunn. Chami. (2009a). Author House. Yahweh and the Gods of Canaan: A Historical Analysis of two Contrasting Faiths. (Eds. (Ed. Ashby. Cambridge University Press. Ringgren. Heinz-Josef. Mubabinge.A. G. Helmer and Fabry. Barnes. 2nd Edition. F. (Ed. The Ancient Egyptian Book of the Dead: The Book of Coming Forth by Day. Facts on File.). The Ancient Egyptian Pyramid Texts. Augustine.BIBLIOGRAPHY Aiyejina.J. Penny Farthing Press. Childs. RSP. Cruzian Mystic Books. Indiana University Press. William. Sumerian Comparative Dictionary. (1990). Sandra T. Africa’s Ogún: Old World and New. (2005). ________________. Berkerie. Eerdmans Publishing. (2000). Hans (1992). Muata (2001). “Esu Elegbara: A Source of an Alter/Native Theory of African Literature and Criticism. Comparative Linguistics: Indo-European and Niger-Congo. J. Clyde W. Athelia Henrietta Press. Brill. Inprint Editions. Afolabi A. (2001). Kosha Anja and Alfred. Gorman. Griffith Institute.).” (Doctoral dissertation). Robert. MOCHA-Versity Press. Griffiths. Ogún. The Origins of Osiris and his Cult. Asar (2009). Fu-Kiau. Naomi. Mark. Scientific and Technological Thinking.. Bantam Books. Joseph E. MOCHA-Versity Press (forth coming). Beyond You and Me: Inspirations and Wisdom for Building Community. Africanisms in American Culture. Chukwunyere. Self-Healing Power and Therapy: Old Teachings from Africa.. Erman. Lawrence Hill Books. Doumbia.Dingemanse. Khenti. Ifa: The Ancient Wisdom.O. The Bakala of North America – The living Suns of Vitality: In Search of a Meaningful Name for African-Americans. Bantam Books. David C. I-V. Halloway. A Concise Dictionary of Middle Egyptian. Gooding. Tweney. Ford. Llewellyn Publications Epega. Alexandra P. Kamalu. (1999). ________________. “The Body in Yorùbá: A Linguistic Study. (2003). Michael E. Person. and Kincannon. (2004).(2005). Unveränderter Nachdruck. Hermann. (1998).. (1999). Indiana University Press. Bunseki. Gnuse. African Fire Philosophy and the Meaning of KMT. (2005). Adama and Doumbia. Robert K. Kimbwandende K. No Other Gods: Emergent Monotheism in Israel. Clyde W. (2006). Dianne D. (1971). (1997). (1980). Karnak House Publishing. (2011). Sheffield Academic Press. R. ________________. (1998). African Cosmology of the Bantu Kongo: Principles of Life and Living. Faulkner. John G. 149 . Charles S. E. University of Leiden. The Hero with an African Face: Mythic Wisdom of Traditional Africa. Adolph and Grapow. Permanent Publishing. The Star of Deep Beginnings: The Genesis of African Science and Technology. Inc. (2010). (1991). Joubert. Ford. (Ed. Ashmolean Museum. The way of the Elders: West African Spirituality and Traditions. Berlin Finch. Glave. Ryan D. Inc. The Hero with an African Face: Mythic Wisdom of Traditional Africa.(Eds). WÖRTERBUCH DER AEGYPTISCHEN SPRACHE im Auftrage der deutschen Akademien hrsg Bd. (2007). (1962). Rooted in the Earth: Reclaiming the African-American Environmental Heritage. Athelia Henrietta Press. 2nd Edition. Divinity & Nature: A Modern View of the Person & the Cosmos in African Thought. Imhotep. Lawrence Erlbaum Associates. ). Allen.1900 C. Ritual. Ra. Memory: Luba Art and the Making of History. and Community. Inc. D. A Jeremy P. Ancient Egyptian & Semitic. Black Arabia and the African Origin of Islam. (1964). Vusamazulu Credo. ________________. JOM International. Obenga. (1984). 150 . Kairn A. Sankore Press. Karnak House Publishing. Wesley (2009). Jacob H. Cambridge. (1976). Murdock. Niemark. Oriental Institute of the University of Chicago. (1993). Maulana and Carruthers. Antonio. Prestel Saakana. Muhammad.M. (1995). (1992).). 1. (1986). Amon Saba (Ed. University of California Press. Readings in Precolonial Central Africa: Texts and Documents. (2009). Magic. Linguistics & Gender Relations. Daystar Press ________________. Karnak House Publishing.Karenga. Somé. Modupe. Karnak Books. (1998). Rescue and Restoration. M. and Initiation in the Life of an African Shaman. Words and Meaning in Yorùbá Religion: Linguistic Connections in Yorùbá. Destiny Books. Prophecies and Mysteries. The Way of the Òrìṣà: Empowering your life through the ancient African religion of Ifa. ________________. (2007). Stellar House Publishing.” Self-published paper. Heinemann Lichtheim. Orbis Books. Yorùbá Names: Their Structure and their Meanings. Oduyoye. Zulu Shaman: Dreams. The Sons of Gods and the Daughters of Men: An Afro-Asiatic Interpretation of Genesis 1-11. Loprieno. Tarcher/Putman Book. A-Team Publishing. African Origins of the Major World Religions. Jeremy P. (Ed. (1995). Malidoma. (2003<1996). (Eds. The Pygmies Were Our Compass: Bantu and Batwa in the History of West Central Africa. Gove Press. (1975). Of Water and the Spirit: Ritual. Philip J. Klieman. Roberts. Miller. ________________. Studies in Semitic and Afro-Asiatic Linguistics Presented to Gene B. (2003). HarperCollins Publishers. Christ in Egypt: The Horus-Jesus Connection. (1994). (1991). Mutwa. 3rd Edition. Ancient Egyptian: A Linguistic Introduction. Gragg. (2007). ________________. Early Times to c. Karnak House Publishing Oduwafun. Indaba My Children: African Folktales. Ancient Egypt & Black Africa: A Student’s Handbook for the Study of Ancient Egypt in Philosophy. (1995). Cynthia L. Ancient Egyptian Literature -. (2001). (1996). The Healing Wisdom Africa: Finding Life Purpose Through nature.E. Let the Ancestors Speak: Removing the Veil of Mysticism from Medu Netcher Vol. Ankh Mi. Tarcher/Putman. “Ewe Orisha: A Treatise on the Role of Plants in Yorùbá Religion. Theophile.Volume I: The Old and Middle Kingdoms. Kemet and the African Worldview: Research.). (2010 – July).biblestudytools. Early Dynastic Egypt. Mark.9171. The Complete Gods and Goddess of Ancient Egypt. The Secret Life of Plants. Flash of the Spirit: African & Afro-American Art & Philosophy. Our Common African Genesis: Evedence from Genetics. (2003).com CiLuba Dictionary. (2001). Toby A. Websites Canaanite Dictionary. Wilkinson.ciyem.923112. Genesi and Ancient History and How Judeo-Christian Mythology Tried to Erase It.html 151 .ancientscripts. Ltd. Christopher. Robert F.html Did Ancient River Channels Guide Humans Out of Africa. West. Vantage Press. Richard H. Archeology. Wilkinson.html Bible Study. First Vintage Books. Routledge.be The Beinlich Egyptian-German Wordlist. Peter and Bird.fitzmuseum. Inc.org/ Yorùbá Online Dictionary Ancient Scripts The Sahara’s Buried Rivers. Thames & Hudson. Ancient Egyptian Dictionary. HarperCollins India.yorubadictionary. (2000<1973).newscientist.com/article/dn14935-did-ancient-river-channels-guide-humans-out-ofafrica. (1983).ugent. Tompkins. Linguistics. Larry (2009).00. Vygus.
https://www.scribd.com/doc/58722498/Passion-of-the-Christ-or-Passion-of-Osiris-The-Kongo-Origins-of-the-Jesus-Myth
CC-MAIN-2017-09
refinedweb
68,528
66.94
20 Analyzing financial data is a critical requirement for any organization. In financial analytics, it is important to understand the basics of financial data and be able to analyze it. In this guide, you will learn the fundamental financial concepts of growth, time value of money, cost of capital, and the capital budgeting methods to evaluate project financing decisions. You will also learn how to compute these measures using Python. The entire world is driven by growth and returns. Companies that manage to maintain their growth rates succeed, while others fall behind. We'll look more closely at this concept, but before that, we'll create a fictitious dataset with pandas using the lines of code below. The first line of code below imports the pandas library, while the second line initializes the list. The third line of code uses the .DataFrame() function to convert this list into a data frame, while the fourth line prints the data. 1 2 3 4 5 6 import pandas as pd dat = [{'2015': 5, '2016':6, '2017':8, '2018':10, '2019':13}, {'2015': 4, '2016':4.8, '2017':6.1, '2018':8, '2019':10},{'2015': 1, '2016':1.2, '2017':1.9, '2018':2, '2019':3},] df = pd.DataFrame(dat, index =['Revenue', 'Expenses', 'Profit']) df Output: 1 2 3 4 5 | | 2015 | 2016 | 2017 | 2018 | 2019 | |---------- |------ |------ |------ |------ |------ | | Revenue | 5 | 6.0 | 8.0 | 10 | 13 | | Expenses | 4 | 4.8 | 6.1 | 8 | 10 | | Profit | 1 | 1.2 | 1.9 | 2 | 3 | The output above is a simplified version of the income statement, but it will serve our purpose of understanding the concept of growth. If we want to calculate the year-on-year growth rate in revenues for 2019, we can use the first line of code below, and the second line will print the resulting data with the new variable. 1 2 df['YoY Growth'] = ((df['2019'] - df['2018'])/df['2018'])*100 df Output: 1 2 3 4 5 | | 2015 | 2016 | 2017 | 2018 | 2019 | YoY Growth | |---------- |------ |------ |------ |------ |------ |------------ | | Revenue | 5 | 6.0 | 8.0 | 10 | 13 | 30.0 | | Expenses | 4 | 4.8 | 6.1 | 8 | 10 | 25.0 | | Profit | 1 | 1.2 | 1.9 | 2 | 3 | 50.0 | From the above output, we can infer that year-on-year Revenue, Expenses and Profit growth in 2019 were 30 percent, 25 percent, and 50 percent, respectively. However, looking at only one year growth might be misleading, which is why it's preferred to look at the compounded annual growth rate (CAGR). CAGR measures mean growth rates of numeric variables over the years and is a better metric for analyzing growth rates. We can calculate CAGR using the code below. 1 2 df['CAGR'] = ((df['2019']/df['2015'])**(1/4.0)-1)*100 df Output: 1 2 3 4 5 | | 2015 | 2016 | 2017 | 2018 | 2019 | YoY Growth | CAGR | |---------- |------ |------ |------ |------ |------ |------------ |----------- | | Revenue | 5 | 6.0 | 8.0 | 10 | 13 | 30.0 | 26.982343 | | Expenses | 4 | 4.8 | 6.1 | 8 | 10 | 25.0 | 25.743343 | | Profit | 1 | 1.2 | 1.9 | 2 | 3 | 50.0 | 31.607401 | We can compare the difference between the one-year growth rate, YoY Growth, and the compounded annual growth rate, CAGR. For example, annual profit growth is 50 percent in 2019, but the CAGR for the years 2015 through 2019 is approximately 32 percent. This difference illustrates the importance of measuring financial performance with CAGR. The growth rates we discussed above can be used for future projections as well. For instance, if we want to project a financial statement for this data three years from now using CAGR, we can easily do so using the lines of code below. 1 2 df['Projection_3year'] = df['2019']*(1+(df['CAGR']/100))**(5) df Output: 1 2 3 4 5 | | 2015 | 2016 | 2017 | 2018 | 2019 | YoY Growth | CAGR | Projection_3year | |---------- |------ |------ |------ |------ |------ |------------ |----------- |------------------ | | Revenue | 5 | 6.0 | 8.0 | 10 | 13 | 30.0 | 26.982343 | 42.920032 | | Expenses | 4 | 4.8 | 6.1 | 8 | 10 | 25.0 | 25.743343 | 31.435836 | | Profit | 1 | 1.2 | 1.9 | 2 | 3 | 50.0 | 31.607401 | 11.844666 | The time value of money is one of the most important concepts in financial investing. The simple rationale is that the present value of money will depreciate with time, which means that one dollar in your pocket today is more than the one dollar in your pocket one year from now because of inflation. Let's illustrate this by calculating the present value of $100 five years from now, assuming inflation of 3 percent, using the lines of code below. We use the pv(rate, nper, pmt, fv) function from the numpy package to calculate the present value of money with a few simple parameters. rate: The rate of return. nper: The timespan of the investment. pmt: The fixed payment at the beginning or end of each period. In our example, this value is zero. fv: The future value of the investment. 1 2 import numpy as np np.pv(rate=0.03, nper=5, pmt=0, fv=100) Output: 1 -86.2608784384164 Similarly, we can calculate the future value of $100 invested for 5 years at a 3 percent annual rate of return using the numpy's .fv(rate, nper, pmt, pv) function, using the lines of code below. 1 np.fv(rate=0.03, nper=5, pmt=0, pv=-100) Output: 1 115.92740743 The future value comes out to be $115. It is important to note that in the above function, we have passed the negative value into the pv parameter. This is because it represents a cash outflow made in the first year and is required for calculating the future value. Capital budgeting is the process in financial management of evaluating potential projects or investments. Construction of a new plant or purchase of a new machinery are examples of capital budgeting projects. There are several methods for evaluating projects. We'll consider the two most popular ones—net present value (NPV) and internal rate of return (IRR)—and their implementation in Python. In simple terms, the net present value (NPV) is the difference between the present value of a project's cash inflows and cash outflows over a period of time. We'll use the numpy function .npv(array of values) to compute the net present value. Let's take four projects with the following cash flows. The discount rate is 5 percent for all these projects. Project 1: Cash outflow of -$1000 in first year, followed by cash inflows of $1000 in second year, $2000 in third year, $3000 in fourth year, and $4000 in the fifth year. Project 2: Cash inflows of $1000 in first year and $2000 in second year, followed by cash outflow of -$1000 in third year, inflow of $3000 in fourth year, and inflow of $4000 in fifth year. Project 3: Cash outflow of -$1000 in first year, followed by cash inflows of $4000 in second year, $3000 in third year, $2000 in fourth year, and $1000 in fifth year. Project 4: Cash outflows of -$1000 and -$2000 in first and third years with cash inflows of $3000 in second year, $2000 in fourth year, and $7000 in fifth year. It's important to note that even though the cash flow pattern is different in these four projects, the absolute project value for each project comes out to be $9000 if we sum all the cash flows. So the question is, which of these projects should be selected? The answer comes from the magical concept of the time value of money, which we'll use to calculate the present value of all the cash flows. The lines of code below perform this task. 1 2 3 4 5 6 7 8 9 10 11 # Project 1 NPV print(np.npv(rate=0.05, values=np.array([-1000, 1000, 2000, 3000, 4000]))) # Project 2 NPV print(np.npv(rate=0.05, values=np.array([1000, 2000, -1000, 3000, 4000]))) # Project 3 NPV print(np.npv(rate=0.05, values=np.array([-1000, 4000, 3000, 2000, 1000]))) # Project 4 NPV print(np.npv(rate=0.05, values=np.array([-1000, 3000, -2000, 2000, 7000]))) Output: 1 2 3 4 5 6 7 7648.76 7880.05 8080.98 7529.67 The above output shows that the NPV is highest for the third project, so with everything else being equal, the third project should get preference over the others. While NPV is useful, another important measure widely used by finance professionals for selecting capital budgeting proposals is the IRR method, as discussed below. Internal rate of return (IRR) is the rate of return at which the net present value of all the cash flows from a project or investment becomes equal to zero. IRR is more intuitive to understand than NPV. We'll use the numpy .irr(array of values) function to compute the internal rate of return. Let's calculate the IRR of all four projects discussed above using the lines of code below. 1 2 3 4 5 6 7 8 9 10 11 # Project 1 IRR print(np.irr(np.array([-1000, 1000, 2000, 3000, 4000]))) # Project 2 IRR print(np.irr(np.array([1000, 2000, -1000, 3000, 4000]))) # Project 3 IRR print(np.irr(np.array([-1000, 4000, 3000, 2000, 1000]))) # Project 4 IRR print(np.irr(np.array([-1000, 3000, -2000, 2000, 7000]))) Output: 1 2 3 4 5 6 7 1.517996077869768 nan 3.732626152716434 1.8474807808952747 The IRR for Project 1, Project 3, and Project 4 is 151 percent, 373 percent, and 185 percent, respectively. The second project does not return any IRR because there is no cash outflow in the first year, which is a requirement for calculating IRR and a limitation of this method. Using the IRR method, we would select Project 3 with the highest IRR of 373 percent. IRR is a good metric to evaluate investment projects. However, what if the rate of return is less than the cost of financing the investment? For instance, if the IRR of a project is ten percent and the cost of capital for funding the project is 12 percent, then it will not be a lucrative idea to invest in that project. That is why it's necessary to compute the weighted average cost of capital. The weighted average cost of capital (WACC) is the rate that a company is expected to pay to all its security holders for financing its assets. The WACC is commonly referred to as the firm's cost of capital and is also influenced by external market factors. The formula for calculating the weighted average cost of capital is WACC = (equity_proportion*equity_cost) + (debt_proportion*debt_cost) * (1 - taxrate) where equity_proportion: The percentage of a company's financing via equity. equity_cost: The cost of equity for the company. debt_proportion: The percentage of a company's financing via debt. debt_cost: The cost of debt for the company. taxrate: The corporate tax rate. Let's calculate the WACC of a company with a 10 percent cost of debt, 15 percent cost of equity, 40 percent debt financing and 60 percent equity financing. The effective corporate tax rate is 30 percent. It's simple to calculate the WACC using the above formula. 1 2 3 4 5 6 7 equity_proportion = 0.60 debt_proportion = 0.40 equity_cost = 0.15 debt_cost = 0.10 taxrate = 0.30 WACC = (equity_proportion*equity_cost) + (debt_proportion*debt_cost) * (1 - taxrate) WACC Output: 1 0.118 The WACC comes out to be 11.8 percent. The simple interpretation of this is that if the IRR of the project is more than the WACC for the company, it makes sense to select the project proposal. In this guide, you learned about the basics of analyzing financial data. You learned the important concepts of growth, CAGR and the time value of money. You also learned several capital budgeting techniques for evaluating investment project proposals. Finally, you learned how to implement these concepts in Python. To learn more about data science using Python, please refer to the following guides. 20
https://www.pluralsight.com/guides/implementing-data-analysis-with-financial-data
CC-MAIN-2020-29
refinedweb
2,005
74.08
Overview Atlassian SourceTree is a free Git and Mercurial client for Windows. Atlassian SourceTree is a free Git and Mercurial client for Mac. Welcome to Tipfy! Tipfy is a small but powerful framework designed specifically for Google App Engine. It is a lot like webapp: from tipfy import RequestHandler, Response class HelloWorldHandler(RequestHandler): def get(self): return Response('Hello, World!') ...but offers a lot of features (own authentication, sessions, i18n etc) and other goodies that webapp misses. Everything in a modular, lightweight way, tuned for App Engine. You use only what you need, when you need. Read the documentation to learn more about it: For questions and comments, join our discussion group: And if you have any issues, open a ticket at Google Code: Installation Read the instructions in project/README.txt or for a more detailed description see installation instructions.
https://bitbucket.org/jeverling/tipfy
CC-MAIN-2016-22
refinedweb
141
57.67
Dynamic Linking Library Functions dlsym(3DL) NAME dlsym - get the address of a symbol in a shared object or executable SYNOPSIS cc [ flag ... ] file ... -ldl [ library ... ] #include <dlfcn.h> void *dlsym(void *handle, const char *name); DESCRIPTION The dlsym() function allows a process to obtain the address of a symbol defined within a shared object or executable. The handle argument is either the value returned from a call to dlopen() or one of the special handles RTLD_DEFAULT, RTLD_NEXT, or RTLD_SELF. The name argument is the symbol's name as a character string. In the case of a handle returned from dlopen(), the corresponding shared object must not have been closed using dlclose(). The dlsym() function searches for the named sym- bol in all shared objects loaded automatically as a result of loading the object referenced by handle. See dlopen(3DL). In the case of the special handle RTLD_DEFAULT, dlsym() searches for the named symbol starting with the first object loaded and proceeding through the list of initial loaded objects, and any global objects obtained with dlopen(3DL), until a match is found. This search follows the default model employed to relocate all objects within the process. In the case of the special handle RTLD_NEXT, dlsym() searches for the named symbol in the objects that were loaded following the object from which the dlsym() call is being made. In the case of the special handle RTLD_SELF, dlsym() searches for the named symbol in the objects that were loaded starting with the object from which the dlsym() call is being made. In the case of RTLD_DEFAULT, RTLD_NEXT, and RTLD_SELF, if the objects being searched have been loaded from dlopen() calls, dlsym() searches the object only if the caller is part of the same dlopen() dependency hierarchy, or if the object was given global search access. See dlopen(3DL) for a discussion of the RTLD_GLOBAL mode. RETURN VALUES If handle does not refer to a valid object opened by dlo- pen(), is not the special handle RTLD_DEFAULT, RTLD_NEXT, or RTLD_SELF, or if the named symbol cannot be found within any SunOS 5.8 Last change: 13 Mar 2000 1 Dynamic Linking Library Functions dlsym(3DL) of the objects associated with handle, dlsym() will return NULL. More detailed diagnostic information is available through dlerror(3DL). EXAMPLES Example 1: Using dlopen() and dlsym() to access a function or data objects. The following example shows how one can use dlopen() and dlsym() to access either function or data objects. For sim- plicity,); Example 2: Using dlsym() to verify that a particular func- tion is defined. The following code fragment shows how dlsym() can be used to verify that a particular function is defined and to call it only if it is. int (*fptr)(); if ((fptr = (int (*)())dlsym(RTLD_DEFAULT, "my_function")) != NULL) { (*fptr)(); } USAGE The dlsym() function is one of a family of functions that give the user direct access to the dynamic linking facili- ties (see Linker and Libraries Guide) and are available to dynamically-linked processes only. ATTRIBUTES See attributes(5) for descriptions of the following attri- butes: SunOS 5.8 Last change: 13 Mar 2000 2 Dynamic Linking Library Functions dlsym(3DL) ____________________________________________________________ | ATTRIBUTE TYPE | ATTRIBUTE VALUE | |_____________________________|_____________________________| | MT-Level | MT-Safe | |_____________________________|_____________________________| SEE ALSO ld(1), dladdr(3DL), dlclose(3DL), dldump(3DL), dlerror(3DL), dlopen(3DL), attributes(5) Linker and Libraries Guide SunOS 5.8 Last change: 13 Mar 2000 3
http://www.manpages.info/sunos/dlsym.3.html
CC-MAIN-2016-44
refinedweb
569
57
NAME getfh, lgetfh - get file handle LIBRARY Standard C Library (libc, -lc) SYNOPSIS #include <sys/param.h> #include <sys/mount.h> int getfh(const char *path, fhandle_t *fhp); int lgetfh(const char *path, fhandle_t *fhp); DESCRIPTION The getfh() system call returns a file handle for the specified file or directory in the file handle pointed to by fhp. The lgetfh() system call is like getfh() except in the case where the named file is a symbolic link, in which case lgetfh() returns information about the link, while getfh() returns information about the file the link references. These system calls are restricted to the superuser. RETURN VALUES Upon successful completion, the value 0 is returned; otherwise the value -1 is returned and the global variable errno is set to indicate the error. ERRORS The getfh() and lgetfgh() system calls fail if one or more of the following are true: [ENOTDIR] A component of the path prefix of path is not a directory. [ENAMETOOLONG] The length of a component of path exceeds 255 characters, or the length of path exceeds 1023 characters. [ENOENT] The file referred to by path does not exist. [EACCES] Search permission is denied for a component of the path prefix of path. [ELOOP] Too many symbolic links were encountered in translating path. [EFAULT] The fhp argument points to an invalid address. [EIO] An I/O error occurred while reading from or writing to the file system. HISTORY The getfh() system call first appeared in 4.4BSD.
http://manpages.ubuntu.com/manpages/maverick/man2/lgetfh.2freebsd.html
CC-MAIN-2015-14
refinedweb
248
62.58
On Thu, Oct 02, 2008 at 12:29:30AM +0800, Tom Stellard wrote: > I think I can improve my script, so that I won't need all of that > preprocessor stuff. However, I had some trouble when trying to > include check.as. There is one check.as file in > testsuite/actionscript.all for the actionscipt files, and a check.as > file in testsuite/actionscript.all/haxe-swf9 for the haxe files. I > thought I was setting the correct include directory here in the > Makefile (located in testsuite/actionscript.all): > > HAXE_DIR = $(srcdir)/haxe-swf9 > HAXE_CPP = $(CPP) -DHAXE -DOUTPUT_VERSION=9 -x c -P -I$(HAXE_DIR) > > But when I actually run the preprocessor, it always uses the check.as > file in the testsuite/actionscript.all directory. That is why I added > the #ifdefs around the #include "check.as" here: > > +#ifdef HAXE > +#include "haxe-swf9/check.as" > +#else > #include "check.as" > +#endif > > Can anyone spot what I am doing wrong? I've no problem with that change in check.as really. I guess that's required because cpp(1) will still look in CWD before moving on ? I'm more concerned with the changes in String.as (the actual testcases). --strk;
http://lists.gnu.org/archive/html/gnash-commit/2008-10/msg00012.html
CC-MAIN-2019-04
refinedweb
197
70.6
Be the first to know about new publications.Follow publisher Unfollow publisher Josh Rattray - Info Spread the word. Share this publication. - Stack Organize your favorites into stacks. - Like Like this publication. 2010 Butler Football Guide 2010 Butler University football guide 2010 Butler Football Contents All-Americans, Butler ...................................... 28 Butler Bowl ....................................................... 6 Butler University ............................................. 40 Coaching Records, Butler............................... 32 Coaching Staff ............................................... 4-6 Head Coach Jeff Voris ................................. 4 Joe Cheshire ................................................ 5 Tim Cooper .................................................. 5 Neil Tabacca................................................. 5 Lettermen, Butler All-Time..................... 36-39 Opponents ................................................. 24-26 Pioneer Football League ................................ 27 Player Profiles ............................................. 7-18 Records, Annual Team Won-Lost ................... 32 Records, Butler Individual.......................... 28-31 Records, Butler Team ..................................... 29 Results, 2009 Game-by-Game ....................... 23 Results, All-Time........................................ 33-35 Roster, 2010 .............................................. 20-21 Schedule, 2010................................................. 1 Statistics, 2009 ............................................... 22 Butler Quick Facts Location: Indianapolis, Ind. 46208 Founded: 1855 Enrollment: 4,200 Nickname: Bulldogs Colors: Blue and White Conference: Pioneer Football League National Affiliation: NCAA Football Championship Subdivision (FCS): 24-21 (4 yrs.) Career Record: 39-55 (9 yrs.) Office Phone: (317) 940-6803 Best Time to Call: 9 a.m.-12 noon Assistant Coaches: Joe Cheshire (DePauw ‘99) Tim Cooper (DePauw ‘97) Nick Tabacca (Ball State ‘04) Athletic Trainer: Chris Tinkey Stadium: Butler Bowl Capacity: 5,647 Surface: Pro Grass (Field Turf) Press Box Phone: (317) 940-9817 2009 Record: 11-1-0 2009 League Record: 7-1 (1st-tie) Lettermen Returning: 40 (20 Off., 18 Def., 2 Spec.) Lettermen Lost: 13 Starters Returning: 17 (9 Off., 8 Def.) All-Time Football Record: 539-417-35/120 years Safety Mark Giacomantonio celebrated Butler’s come-from-behind Homecoming victory over San Diego. The dramatic win helped the Bulldogs compile a per perfect 7-0 record in the Butler Bowl in 2009. 2010 BUTLER FOOTBALL SCHEDULE Game Time at Site Date Opponent Game Site City Sept. 4 At Albion Sprankle-Sprandel Stadium Albion, Mich. 1 p.m. Sept. 11 At Youngstown State Stambaugh Stadium Youngstown, Ohio 6 p.m. Sept. 18 TAYLOR Butler Bowl Indianapolis, Ind. 1 p.m. Sept. 25 At San Diego* Torero Stadium San Diego, Calif. 6 p.m. Oct. 2 CAMPBELL* Butler Bowl Indianapolis, Ind. noon Oct. 9 At Davidson* Richardson Stadium Davidson, N.C. noon Oct. 16 DAYTON* Butler Bowl Indianapolis, Ind. 1 p.m. Oct. 23 MOREHEAD STATE*(HC) Butler Bowl Indianapolis, Ind. 1 p.m. Oct. 30 At Valparaiso* Brown Field Valparaiso, Ind. 1 p.m. Nov. 6 JACKSONVILLE* Butler Bowl Indianapolis, Ind. noon Nov. 13 At Drake* Drake Stadium Des Moines, Iowa 1 p.m. *--Pioneer Football League game. (HC)--Homecoming. CREDITS: Butler’s 2010 Football Guide was designed and edited by Jim McGrath, Butler sports information director. Photography by Butler photographer Brent Smith. Cover design by the Butler Publications Office. For the latest in Butler sports information, visit: 1 2010 Butler Football Wide receiver Zach Watkins led the Pioneer Football League and set a Butler record with 78 receptions in 2009. Bulldogs Seek Repeat Performance field goal on the final play of the game! The Bulldogs, who were picked fifth in the league preseason poll, went on to capture a PFL co-championship and advance to the Gridiron Classic, where they defeated Central Connecticut State for their first postseason victory. Butler’s final 11-1 record was the best in school history! This season, fueled by last year’s success, 17 returning starters and a brand new 5,647-seat stadium, the Bulldogs are eager to make a repeat run. But head coach Jeff Voris, entering his fifth first finished field at different times and at different spots, while still training him as a quarterback.” “We definitelyleading receiver last year with 38 catches for 493 yards, while Koopman had 24 receptions for 258 yards. The backfield is well-stocked with Gray, who finished backfield since I’ve been here,” said Voris. “Again, we have to find ways to get them all on the field.” Defensive end Grant Hunter finished 11th in the NCAA Division I FCS in quarterback sacks, despite missing three games with injury. 2010 Butler Football Quarterback Andrew Huck compiled the third-highest passing total in Butler history last seasaon.American, Hunter finished finishedleading tackler Spencer Summerville. Caldicott, Butler’s leading tackler last season, anchors a linebacker corps that includes returning regular Andrew Cottrell and top 2009 reserves Jordan Ridley and William Lamar. Caldicott finished with 70 tackles last season, one more than Guggenberger and two more than Summerville. He also led the team with three forced fumbles and tied for second on the squad with three pass interceptions. Cottrell, playing his first season with the Bulldogs finished fifth final season with seven career pass interceptions and more than 100 career tackles. Giacomantonio, Butler’s fourth-leading tackler last season, led the PFL in pass interceptions in 2009 with five in nine games. Junior Jack McKenna, named Butler’s Most Improved Defensive Player last season, returns at cornerback, where he started all 12 games in 2009 and tied Dombart with three pass thefts. He also finished field position. Junior David Lang took over as the team’s placekicker, and he finished as the Butler’s third-leading scorer with 54 points. He booted seven field goals, including a pair of game-winners, and wound up third in the PFL in kick scoring. Koopman returns as the squad’s top kick returner, averaging 22.7 yards per return last season, while Dombart led the Bulldogs in punt returns. Tailback Scott Gray led the Bulldogs and ranked third in the PFL with a career-best 868 rushing yards in 2009. 3 Head Coach Jeff Voris In just four seasons, Jeff Voris has taken Butler from an 0-11 record to the most successful season in school history! Voris, who’s improved the Bulldogs’ win total in each of his four seasons, took a Butler team picked fifth in the Pioneer Football League preseason poll, and guided it to an 11-1 campaign, a tie for the PFL championship and Butler’s first-ever post-season victory. The 11 wins in 2009 set a Butler single season record and tied the PFL single season mark, and the league title was Butler’s first since 1994. Butler also tied the PFL record for league wins (7), while setting a record for home wins (7) and completing an unbeaten Butler Bowl slate. The Buldogs. To no one’s surprise, Voris was named the PFL Coach of the Year. The historic 2009 Butler campaign came one year after Voris guided the Bulldogs to a 6-5 record, the program’s first winning season in more than a decade! The six wins were Butler’s most since 1997 and the second-most since the Bulldogs moved to Division I in 1993. And Butler’s ffour 4 The Jeff Voris File Name: Jeff Voris Age: 43 Birthdate: 8/27/67 Wife: Julie Children: Jenna (14), Josie (12), Jessy 2009 Career School Carroll Carroll Carroll Carroll Carroll Butler Butler Butler Butler Collegiate Head Coaching Record Won Lost Pct. 2 7 .222 1 9 .100 3 7 .300 3 7 .300 6 4 .600 3 8 .273 4 7 .364 6 5 .545 11 1 .917 39 55 .415 (14), Josie (12) and Jessy (9). Assistant Coaches Joe Cheshire Co-Defensive Coordi Coordinator Joe Cheshire is in his sixth season as a full-time assistant coach, following one season with the Bulldogs as a parttime assistant. He handles the team’s linebackers, while also serving as Butler’s recruiting coordinator, and this season, he’s taking on the additional duties of Co-Defensive. Nick Tabacca Offensive Line Nick Tabacca, who served the past three seasons on the football staff at Defiance College, was named an assistant football coach at Butler in July, 2010. He took over as offensive line coach for the Bulldogs, filling the spot of departed coach and offensive coordinator Frank Smith. Tabacca was the offensive line coach for all three of his seasons at Defiance, and he spent the past two campaigns as offensive coordinator for the Yellow Jackets. He directed an offensive unit that led the Heartland Collegiate Athletic Conference in fewest turnovers for two straight years. During his three seasons at the Ohio school, Defiance had seven AllConference Defiance in 2007. He received a B.S. degree from Ball State in 2004 and earned his masters degree in business administration from Ball State in 2006. Tim Cooper Co-Defensive Coordi Coordinator Former Miami (Ohio) and Carroll College football assistant Tim Cooper was named an assistant football coach at Butler in March, 2010. He’s working with Butler’s defensive secondary and serves as a Co-Defensive Coordinator. Cooper was. David Kenney Defensive Tackles David Kenney is in his fifth season on Butler’s coaching staff, and he has responsibility for the Bulldogs’ defensive line. He spent a portion of the 2010 preseason working with the Indianapolis Colts’ defensive unit under the Bill Walsh Minority Internship Program. Kenney (17), David (15), Khirra (10) and Dakari (9). He’s a member of Kappa Alpha Psi Fraternity, Inc. Nick Anderson Cornerback Cornerbacks Nick Anderson is in his fourth season with the Bulldogs, taking over this year as cornerbacks’ coach. He’s also coached wide receivers for the Bulldogs. Anderson Receiver Receivers Danny Sears, a former standout wide receiver at Franklin College, is in his fourth season at Butler. He’s working with the team’s receivers this year, after coaching cornerbacks in 2009. Sears was a three-time, first team All-Heartland Conference wide receiver (2004-06) and a two-time conference Special Teams MVP (2004 & 2006) as a punt and kick returner. He finished his collegiate career with more than 2,000 receiving yards and 20 career touchdown catches. The four-year letterwinner for the Grizzlies compiled over 5,000 all-purpose yards, 2003-06. Chris Davis Running Backs ack acks Chris Davis, who earned two varsity letters ers as an of offensive lineman with the Bulldogs, is in his third season on the Butler staff. He’s coaching Butler’s running backs, after working with the squad’s tight ends last season.. 5 Assistant Coaches/Staff Rob Noel Defensive Ends Matt Walker Tight Ends Rob Noel, a four-year letterwinner with the Bulldogs (2005-08), is in his second final 22 games. He earned AllLeague honors as an offensive lineman, and he was a two-time Academic All-League player. Noel earned his degree from Butler in 2009. Matt Walker, former head coach at DePauw University, is in his first season on Butler’s staff. He has responsibility. SUPPORT STAFF Butler Bowl Thorn Murphy Student Assistant Offense 6 Grant Lewis Student Assistant Defense Chris Tinkey Football Athletic Trainer Lester Burris Student Video Coordinator Stephen Blowers Student Manager Ted Pajakowski Student Manager John Harding Equipment Manager Jennifer Johnson Equipment Room Jim Peal Strength/Conditioning Coordinator The Butler Bowl, home to Butler football for eight decades, received a major face-lift for the 2010 season. Located on the northeast side of Butler’s campus just east of storied Hinkle Fieldhouse, the Butler Bowl project took a major step forward with the addition of new bleacher and chairback seats on the west side, visitor’s seating on the east side and a new brick press box stretching between the 30-yard lines. The new seats boosted the stadium capacity to 5,647, and the new press box added main level areas for media, radio, coaches and boosters as well as an upper level video area and observation deck. The renovation of the Butler Bowl, part of a plan to build new student housing along the east side of the stadium,. 2010 Bulldogs 46 Returning Letterwinners 33 BILL BORK 6-1, 225, Jr. TIGHT END Dyer, Ind. (Lake Central) 2009 Career RUSH 0 0 0 YDS. 0 0 0 14 AVG. 0.0 0.0 0.0 TD 0 0 0 REC. 2 2 4 NICK CALDICOTT 6-1, 225, Jr. YDS. 11 11 22 TD 0 0 0 YEAR 2008 2009 Career ST 9 39 48 AT 5 31 36 TT 14 70 84 FOR LOSS SACKS 0-0 0 8-22 1-7 8-22 1-7 95 TAYLOR CLARKSON DEFENSIVE TACKLE 6-2, 275, So. Fortville, Ind. (Mt. Vernon) YEAR 2009 ST 2 AT 0 TT 2 FOR LOSS SACKS 0-0 0 YEAR 2007 2008 2009 Career ST 0 6 15 21 FR 0 AT 0 5 11 16 TT 0 11 26 37 FOR LOSS SACKS 0-0 0 4-20 1 1.5-3 1 4.5-23 2 FR 0 0 0 0 15 ANDREW COTTRELL LINEBACKER 5-11, 215, Jr. West Chester, Ohio (Lakota West/Grand Valley State) 2009: Played in 10 games, including eight as a starter... finished fifth on the squad in tackles...wound up fourth on the team in assisted tackles...recorded a career-high six tackles in five different games...named to the PFL Academic Honor Roll... YEAR 2009 ST 16 FR 0 1 1. DEFENSIVE END Winter Park, Fla. (Winter Park) 2009: Saw action in all 12 games, including seven as a starter... recorded a career-high five tackles in Butler’s victory over Davidson...had a quarterback sack in Butler’s season-opening win over Albion...named to the PFL Academic Honor Roll...2008: Played in all 11 games...tied for fifth on the team in tackles for loss...had a season-high four tackles, including two tackles for loss, at Campbell...recorded a quarterback sack at Valparaiso...named to the PFL Academic Honor Roll...2007: Played in three varsity games in initial season at Butler...named to the Pioneer Football League Academic Honor Roll...Personal: Economics major...born 4/8/88...son of Steve and Lynne Cosler...High School: Three-year starter as a defensive lineman at Winter Park High School...team captain and Most Outstanding Defensive Lineman as a senior...earned All-Metro honors in 2006...ranked third in central Florida with 12 quarterback sacks as a senior...helped high school team to a 12-2 record and a state runner-up finish as a junior...earned three varsity football letters...High School Coach: Larry Gergley. LINEBACKER Oak Park, Ill. (Fenwick) 2009: Named Butler’s Defensive MVP in 2009...second team All-PFL...starter in all 12 games...led the Bulldogs in tackles (70) and solo tackles (39)...topped the squad in forced fumbles (3)...tied for second on the team in pass interceptions with three...recorded a career-high 11 tackles in Butler’s victory over Davidson...had a career-best eight solo tackles against Davidson...posted eight tackles against Central Connecticut State in the Gridiron Classic...named to the PFL Academic Honor Roll...2008: Saw action in five games...posted a season-high four tackles against both Franklin and Drake...had a season-best four solo tackles against Franklin...named to the PFL Academic Honor Roll...2007: Did not see varsity action for the Bulldogs...Personal: Management Information Systems. ROB COSLER 6-1, 230, Sr. AT 27 9 TT 43 FOR LOSS SACKS 2.5-4 0 ANDY DAUCH 5-10, 180, Jr. FR 0 DEFENSIVE BACK Bloomfield Hills, Mich. (Lahser) 2009: Saw action in all 12 games...had a career 2009 Career ST 3 5 8 AT 1 5 6 TT 4 10 14 FOR LOSS 0-0 0-0 0-0 INT. 0-0 1-19 1-19 BREAKUPS 0 1 1 7 2010 Bulldogs 36 STEVEN DEPOSITAR 5-10, 195, So. 2009: Played in all 12 games...finished fourth on the team in rushing...had a career-best 50 rushing yards in Butler’s victory at Franklin, including a career-long 41-yard carry... rushed for 47 yards against Hanover...named to the PFL Academic Honor Roll...2008: Didn’t see varsity action in his first season with the Bulldogs...named to the PFL Academic Honor Roll...Personal: Finance. YEAR 2009 RUSH 19 YDS. 130 28 AVG. 6.8 TD 0 REC. 0 TADD DOMBART 5-10, 185, Sr. YDS. 0 TD 0 DEFENSIVE BACK Cincinnati, Ohio (Lakota West) 2009: Played in 11 games, including 10 as a starter...finished second on the team in pass breakups (3) and fourth in pass interceptions (2)...led the Bulldogs and ranked ninth in the PFL in punt returns...had a season-high five tackles against both Drake and Central Connecticut State..high: Finance 2009 Career ST 31 32 20 83 AT 8 10 10 28 TT 39 42 30 111 FOR LOSS 1-1 0.5-2 2.5-4 4-7 INT. 3-47 2-35 2-9 7-91 BREAKUPS 3 5 3 11 Bulldogs Picked Second In PFL Preseason Poll Butler, coming off a Pioneer Football League co-championship, was tabbed second in the PFL’s 2010 preseason coaches’ poll. Ten-time PFL champion Dayton was picked as the coaches’ preseason favorite. Dayton and Jacksonville, chosen third, each recieved four first place votes in the poll, while the Bulldogs received two. Picked in order behind the top three teams were Drake, San Diego, Marist, Davidson, Campbell, Morehead State and Valparaiso. 8 4 RUNNING BACK Mishawaka, Ind. (Penn) MATTHEW FOOR 5-11, 205, So. DEFENSIVE BACK West Chester, Ohio (Lakota West) 2009: Played in 10 games...had a career-high three tackles in Butler’s victory over Central Connecticut State...named to the PFL Academic Honor Roll...2008: Did not see varsity action in first season with the Bulldogs...Personal: Finance. YEAR 2009 ST 7 AT 6 TT 13 FOR LOSS 0-0 INT. 0-0 16 MARK GIACOMANTONIO 5-11, 190, Jr. BREAKUPS 0 DEF. BACK Carmel, Ind. (Carmel) 2009: Honorable mention All-PFL...saw action in nine game, all as a starter...led the PFL and ranked ninth in the NCAA Division I FCS in pass interceptions...finished fourth on the team in tackles...had a career-high 10 tackles in Butler’s victory over Central Connecticut State...named to the 2009 ESPN The Magazine Academic All-District Team and picked first team Academic All-PFL...named to the PFL Academic Honor Roll...2008: Played in all 11 games...recorded a season/Criminology 2009 Career ST 12 30 42 AT 9 15 24 64 TT 21 45 66 FOR LOSS 0.5-2 4-11 4.5-13 DONNIE GILMORE 6-3, 295, Sr. INT. 1-30 5-45 6-75 BREAKUPS 0 2 2 OFFENSIVE LINE Westphalia, Ind. (North Knox) 2009: First team All-PFL offensive lineman...Butler “Battle of the Trenches” Award winner...starter in all 12 games... named to the PFL Academic Honor Roll..: Math/Mechanical Engineering major...born 2/12/88...son of Don and Patricia Gilmore...High School: All-State lineman at North Knox High School...twoway starter for high school squad...all-conference and all-area performer...received Academic All-Area recognition...helped team to a 7-2 record as a senior...earned four varsity football letters...High School Coach: Shawn McDowell. 2010 Bulldogs 22 SCOTT GRAY 5-8, 190, Sr. 2009: First team All-PFL running back...starting tailback in all 12 games...led the Bulldogs and ranked third in the PFL in rushing... finished fifth on the team in scoring...rushed for a season-high 117 yards against Davidson, including a season-long 48-yard rush...had 113 rushing yards against San Diego...rushed for 109 yards against Dayton...named first team Academic All-PFL and chosen to the PFL Academic Honor Roll.. 2009 Total RUSH 141 25 169 335 YDS. 547 101 868 1516 5 AVG. 3.9 4.0 5.1 4.5 TD 15 2 5 22 REC. 6 0 13 19 ARTIS HAILEY III 6-0, 205, So. YDS. 12 0 134 146 TD 0 0 1 1 LINEBACKER Hammond, Ind. (Hammond) 2009: Played in all 12 games as a running back...rushed for a career-best 29 yards in Butler’s victory over Hanover...had a career-best two pass receptions at Franklin...2008: Didn’t see varsity action in his first season with the Bulldogs...earned Butler’s Offensive “Unsung Hero” Award (Scout Team Player of the Year)...Personal: Computer Engineering/Economics. YEAR 2009 RUSH 13 YDS. 65 AVG. 5.0 TD 0 2 RUNNING BACK Lowell, Ind. (Lowell/Ball State) REC. 3 YDS. 4 TD 0 Homework The Bulldogs set a school record with seven victories at home in 2009. It was Butler’s first unbeaten home record as an NCAA Division I program, and the Bulldogs’ first undefeated mark in the Butler Bowl since 1992. The Bulldogs were the lone team in the Pioneer Football League to go unbeaten at home in 2009. Butler’s all-time record in the Butler Bowl is 257-139-11. STUART HARVEY 6-1, 190, Jr. WIDE RECEIVER Carmel, Ind. (Westfield) 2009: Did not see varsity action...named to the PFL Academic Honor Roll...2008: Saw action in all 11 games...finished fifth on the team in receptions and receiving yards...caught a careerhigh four passes, including a seven-yard touchdown reception, against Missouri S & T...had a career-best 32 receiving yards against both Missouri S & T and Campbell...2007: Did not see varsity action in first season with the Bulldogs...Personal: Philosophy/Education, Jr. Bedford, Ind. (Bedford-North Lawrence) 2009: Played in 10 games...finished third on the squad in rushing...wound up seventh in pass receptions...rushed for a season-high 60 yards against Drake...scored a pair of rushing touchdowns against Valparaiso...named to the PFL Academic Honor Roll..: Media Arts 2009 Career RUSH 64 55 120 YDS. 352 346 698 53 AVG. 5.3 6.3 5.8 TD 5 4 9 ROB HOBSON 6-4, 280, Sr. REC. 12 9 21 YDS. 107 50 157 TD 1 0 1 OFFENSIVE LINE Arcadia, Ind. (Sheridan) 2009: Starting center in all 12 Butler games...2008: Saw action in three games...helped the Bulldogs lead the PFL in total offense and rank second in both rushing offense and scoring offense...2007: Did not see varsity action in his initial season at Butler...Personal: Marketing/Digital Media. 9 2010 Bulldogs 10 ANDREW HUCK 6-2, 190, Jr. QUARTERBACK Bloomington, Ind. (Bloomington North) 2009: Named Butler’s Offensive Most Valuable Player and Most Improved Offensive Player...earned honorable mention All-PFL honors...selected as the Most Valuable Player of the 2009 Gridiron Classic...four-time PFL Offensive Player of the Week...named to the College Sporting News National AllStar List on Nov. 9, following a five-touchdown performance at Dayton...starting quarterback in all 12 games...posted the third-highest single season passing total (2,454 yards) in Butler football history...eighth quarterback in Butler history to pass for more than 2,000 yards in a season...finished with the third-highest total for touchdown passes (21) in Butler history...led the PFL in pass completions (233) and pass attempts (371) and finished second in the league in passing yards...wound up third in the league in passing average (204.5)...finished second in the PFL and 27th in the NCAA Division I FCS in total offense (2,921)... ranked 34th in the NCAA in passing efficiency (131.11)...had the league’s highest single game pass completion percentage in 2009 with a 13 of 15 (.867) performance against Hanover...recorded a league-high 33 pass completions in Butler’s victory over Franklin...completed 20 or more passes in six of Butler’s 12 games...threw for a career-high 327 yards in Butler’s victory over Albion...tied Butler’s single game record with a career-best five touchdown passes in the Albion win...had a career-high 395 yards in total offense against Albion...completed 33 of 48 passes (with no interceptions) and threw for 316 yards in Butler’s win at Franklin...tossed four touchdown passes each against Franklin and Morehead State...threw 14 touchdown passes in his first four collegiate starts...ran for a career-high three touchdowns and threw two more against Dayton...completed 15 of 21 passes for 182 yards and rushed for three touchdowns in Butler’s Gridiron Classic victory over Central Connecticut State...rushed for a career-best 81 yards on a career-high 16 carries against Davidson...tied for the team scoring lead and tied for fifth in the PFL in scoring with 10 rushing touchdowns...named first team Academic All-PFL... chosen to the PFL Academic Honor Roll... Year 2008 2009 Total ATT. 1 371 372 COMP. 0 233 233 PCT. .000 .628 .626 98 INT. 0 11 11 TD 0 21 21 YDS. RUSH 0 5 2454 102 2454 107 YDS. AVG. TD 55 11.0 1 467 4.6 10 522 4.9 11 GRANT HUNTER DEFENSIVE END 6-3, 235, Jr. Liberty Twp., Ohio (Lakota West/Robert Morris) 2009: Second team Academic All-America...named second team All-PFL...starting defensive end in nine games...missed three games with illness...finished second in the PFL and 11th in the NCAA Division I FCS in quarterback sacks...led the Bulldogs in tackles for loss and quarterback sacks...finished eighth on the team in total tackles...had a season-high seven tackles against Morehead State and Valparaiso...recorded three quarterback sacks against Morehead State...tied for third on the team in pass breakups (3)...named first team Academic All-PFL, chosen to the ESPN The Magazine Academic All-District squad and selected to the PFL Academic Honor Roll..-high eight tackles, including four quarterback sacks and a career-high six 10: Criminology 2009 Career ST 27 19 46 AT 15 20 35 TT 42 39 81 FOR LOSS SACKS 18-109 14-97 11.5-52 7.5-42 29.5-161 21.5-139 12 MATT KOBLI 6-3, 220, Sr. FR 1 0 1 QUARTERBACK Whiting, Ind. (Whiting) 2009: Sat out the 2009high 48 pass attempts against Franklin...had a career-high 350 total offense yards against Davidson...passed for more than 200 yards in eight of 11 games...had at least 21 pass completions in eight different games...rushed for more/Crim Preseason All-Americans Butler’s Grant Hunter and Zach Watkins each were named to a Preseason All-America Team for 2010. Hunter was chosen as a third team defensive end on the College Sporting News FCS Preseason All-America Team, while Watkins was named a third team wide receiver on The Sports Network Preseason All-America Team. 2010 Bulldogs 1 JORDAN KOOPMAN 5-8, 185, Jr. 2009: Honorable mention All-PFL return specialist...led the Bulldogs and ranked second in the PFL in kick return average (22.7)...ranked fifth in the league in all-purpose yards (1,122)... finished fourth on the team in receiving...had a career-high 146 yards on six kickoff returns at Jacksonville...recorded a career-high six receptions for a career-high 77 yards and a touchdown at Franklin...2008: Saw action in eight games... led the team in kickoff returns and kick return yards...had a career-best six kickoff returns for a season-high 112 yards against Jacksonville...finished eighth on the team in receptions...caught a season-high four passes for a season-high 48 yards against Morehead State...Personal: Exercise Science 2009 Career REC. 11 24 35 51 YDS. 77 258 335 AVG. 7.0 10.8 9.6 TD 1 1 2 KR 20 31 51 ROBERT KOTEFF 6-0, 220, Jr. YDS. 385 703 1088 AVG. 19.2 22.7 21.3 ST 3 0 3 AT 1 1 2 TT 4 1 5 FOR LOSS SACKS 0.5-0 0 0-0 0 0.5-0 0 FR 0 0 0 YEAR 2009 WILLIAM LAMAR 5-10, 195, So. LINEBACKER Oak Park, Ill. (Oak Park-River Forest) 2009: Saw action in all 12 games...had a career-high four tackles against Valparaiso and Davidson...posted a career-high three solo tackles against Jacksonville...had a blocked punt against Morehead State...2008: Did not see varsity action in his initial season at Butler...named to the PFL Academic Honor Roll...Personal: Mechanical-Conference performer...earned two varsity football letters... High School Coach: Jim Nudera. YEAR 2009 ST 10 AT 14 TT 24 FOR LOSS SACKS 1-2 0 FR 0 PAT 33-43 7 PLACEKICKER/PUNTER Lowell, Ind. (Lowell) FG 7-11 LONG 39 PTS. 54 JEFF LARSEN 6-2, 190, Jr. WIDE RECEIVER Chicago, Ill. (Notre Dame) 2009: Saw action in all 12 games...tied for fifth on the squad in receptions...had a career-high four catches for a career-best 42 yards against Jacksonville...recorded three receptions each against Franklin and Campbell...named to the PFL Academic Honor Roll...2008: Did play in any games...named to the PFL Academic Honor Roll...2007: Did not see varsity action... named to the Pioneer Football League Academic Honor Roll...Personal: Finance. YEAR 2009 54 DAVID LANG 6-0, 215, Jr. 2009: Saw action in all 12 games as Butler’s placekicker... two-time PFL Special Teams Player of the Week...wound up third on the team and tenth in the PFL in scoring...third in the PFL in kick scoring...tied for fifth in the league in field goal percentage and finished seventh in the circuit in field goals... kicked a game-winning, 37-yard field goal with 0:01.9 left to lift the Bulldogs past San Diego, 25-24...hit a 27-yard field goal with one second remaining to give Butler a 20-17 victory over Drake and a share of the PFL season championship...had a career-long 39-yard field goal against San Diego...was perfect on six PAT attempts against Albion and Hanover...named first team Academic All-PFL...2008: Played in all 11 games, handling team kickoff chores...averaged 54.7 yards on 57 kickoffs...had one touchback...named to the PFL Academic Honor Roll...Personal: Exercise Science. LINEBACKER Lexington, Ky. (Paul L. Dunbar) 2009: Played in 10 games...named to the PFL Academic Honor Roll... YEAR 2008 2009 Career 35 WIDE RECEIVER Chula Vista, Calif. (Eastlake) REC. 13 72 YDS. 112 AVG. 8.6 PETE MATTINGLY 6-5, 290, Jr. TD 0 OFFENSIVE GUARD Zionsville, Ind. (Cathedral). 11 2010 Bulldogs 81 EDDIE McHALE 6-2, 190, Sr. 2009: Played in all 12 games, including 11 as a starter... finished third on the squad and tenth in the PFL in receiving yards per game...third on the team in receptions...matched his career-high with seven catches at Franklin...had five receptions for a season-high 83 yards against Drake...2008: Saw action in 10 games...finished third on the team in receptions, second in receiving yards, and tied for third in scoring...had a career-high seven catches for 63 yards and a touchdown against Albion...caught four passes for a career-high 100 yards and two touchdowns at Valparaiso...had a career-long 54-yard reception for a TD at Valparaiso...grabbed a 19-yard TD pass against Morhead State...2007: Played in 10 games in initial season with the Bulldogs...had a season-best three catches at Drake...transfer from Ohio University (didn’t play football at Ohio)...Personal: Biology major...born 9/22/87...son of Kevin and Teri McHale...High School: Two-sport athlete at St. Xavier High School...earned three varsity football letters as T a wide receiver...had 35 receptions as a senior...helped lead St. Xavier to a two-year record of 26-1, two conference titles and a state championship, 2004-05...earned three letters in track...High School Coach: Steve Specht. YEAR 2007 2008 2009 Total REC. 9 30 38 77 25 YDS. 53 366 493 912 AVG. 5.9 12.2 13.0 11.8 TD 0 6 3 9 JACK McKENNA 6-0, 180, Jr. ST 2 31 33 AT 6 11 17 TT 8 42 50 FOR LOSS 0-0 1-1 1-1 47 BOB OLSZEWSKI 6-1, 210, Sr. INT. 0-0 3-30 3-30 BREAKUPS 0 5 5 LINEBACKER LaGrange, Ill. (Lyons Township) 2009: Named Butler’s “Unsung Hero” (Scout Team Player of the Year) on defense...played in all 12 games...recorded first career tackle at Franklin... YEAR 2009 12 ST 1 AT 0 TT 1 FOR LOSS SACKS 0-0 0-0 FR 0 JEFF POSS 5-8, 230, Jr. DEFENSIVE END Salt Lake City, Utah (Scripps Ranch, Calif.) 2009: Saw action in all 12 games, including 10 as a starter... finished seventh on the team in tackles...wound up 15th in the PFL in tackles among defensive linemen...fifth on the team in quarterback sacks...had a season-high six tackles against both Dayton and Drake...named to the PFL Academic Honor Roll.. 2009 Career ST 12 24 36 DEFENSIVE BACK Elmwood Park, Ill. (Rochester) 2009: Named Butler’s Most Improved Player...starter in all 12 games...tied for second on the team in pass interceptions... led the team in pass breakups...tied for 10th in the PFL in passes defended...finished sixth on the team in tackles... had a pass interception on the final play of the 2009 Gridiron Classic to seal Butler’s first postseason victory...recorded a career-high seven tackles, all solo, and a pass theft in Butler’s victory over Dayton...named to the PFL Academic Honor Roll...2008: Saw action in all 11 games...had a season-high three tackles at Campbell...named to the PFL Academic Honor Roll...2007: Did not see varsity action.. 2009 Career 52 WIDE RECEIVER Cincinnati, Ohio (St. Xavier/Ohio) AT 8 17 25 TT 20 41 61 FOR LOSS SACKS 6.5-17 1-8 7.5-19 3-13 14-36 4-21 59 JORDAN RIDLEY 5-11, 230, So. FR 0 0 0 LINEBACKER Indianapolis, Ind. (Lawrence Central) 2009: Saw action in all 12 games...finished 10th on the squad in tackles...second on the team in tackles for loss and tied for third on the squad in quarterback sacks...had a careerStar Game...two-time team captain...four-year Scholar-Athlete...four-year football letterwinner...High School Coach: Jayson West. YEAR 2009 ST 20 AT 16 TT 36 FOR LOSS SACKS 8.5-35 4-22 76 RYAN SECRIST 6-4, 275, Sr. FR 1 OFFENSIVE LINE Rochester, Ind. (Rochester) 2009: Missed the entire season with injury...2008: Butler’s starting center in the first five games, before being sidelined. 2010 Bulldogs 97 TYLER SKAGGS 6-2, 240, Sr. 2009: Saw action in 10 games, including seven as a starter...first collegiate tackle was a tackle for loss against Hanover... YEAR 2009 ST 1 AT 3 TT 4 57 DEFENSIVE LINE Winona Lake, Ind. (Warsaw) FOR LOSS SACKS 0.5-1 0-0 FR 0 MIKE STANIEWICZ 6-5, 290, Sr. LOGAN SULLIVAN 6-0, 195, Jr. AT 3 16 19 TT 8 38 46 FOR LOSS 0.5-1 1-1 1.5-2 TT 20 29 49 FOR LOSS SACKS 1.5-3 0 4-14 2-8 5.5-17 2-8 55 JACE TENNANT 6-1, 215, So. FR 0 0 0 DEFENSIVE END Champaign, Ill. (Central) ST 10 AT 6 TT 16 FOR LOSS SACKS 4.5-19 5-17 90 LARRY THOMAS 5-10, 250, Jr. FR 0 DEFENSIVE BACK Winston-Salem, N.C. (Horizon, Ariz.) 2009: Played in all 12 games, including four as a starter... finished ninth on the team in tackles...second on the squad in pass breakups...had a career-high seven tackles against Valparaiso...came up with first career pass interception in Butler’s title-clinching victory over Drake...scored two points on a PAT run at Morehead State to tie the game and send it into overtime...2008: Saw action in all 11 games...had a season...three-time AIA Scholar Athlete Award winner...earned three varsity football letters...High School Coach: Steve Casey. ST 5 22 27 AT 13 20 33 2009: Played in all 12 games...finished second on the team in quarterback sacks...ranked 19th in the PFL in sacks...... recorded first collegiate sack at Franklin...had a pair of quarterback sacks against Hanover...had a career-high three tackles against Albion and Franklin...named to the PFLAcademic Honor Roll...Personal: Finance major...born 7/16/10...son of Mike and Judy Tennant...High School:. YEAR 2009 YEAR 2008 2009 Career ST 7 9 16 OFFENSIVE TACKLE Hebron, Ind. (Lowell) 2009: Second team All-PFL offensive lineman...starting offensive tackle in all 12 games...helped the Bulldogs lead the PFL in total offense and rank second in both rushing offense and scoring offense...named second team Academic All-PFL... selected to the PFL Academic Honor Roll..: Finance major...born 3/12/89...son of April Staniewicz...High School: Two-Conference performer...earned four varsity football letters...High School Coach: Kirk Kennedy. 6 DEFENSIVE TACKLE Sammamish, Wash. (Eastlake) 2009: Saw action in all 12 games, including four as a starter... recorded a career-high five tackles against San Diego...2008: Played in all 11 games...had a season-high four tackles against Morehead State and Jacksonville...Personal: MIS and Marketing 2009 Career 78 ROSS TEARE 6-1, 290, Jr. INT. 0-0 1-10 1-10 BREAKUPS 0 4 4 DEFENSIVE TACKLE West Chester, Ohio (St. Xavier) 2009: Saw action in 11 games...finished sixth on the team in tackles for loss...recorded a career-high five tackles in Butler’s Gridiron Classic victory over Central Connecticut State...had four tackles, including a quarterback sack, at Campbell...2008: Played in all 11 games...had a season-high two tackles in five different games...named to the PFL Academic Honor Roll... Personal: Economics 2009 Career ST 5 8 13 AT 8 9 17 TT 13 17 30 FOR LOSS SACKS 2-18 2-18 5.5-18 1-2 7.5-36 3-20 FR 0 0 0 13 2010 Bulldogs 94 CARTER WALLEY 6-3, 245, So. 2009: Played in all 12 games, including 11 as a starter...scored first collegiate touchdown on a 16-yard pass for Butler’s final score (and eventual game-winner) at Dayton...had a career-high two receptions at Dayton...2008: Did not see varsity action in his first season with the Bulldogs...Personal: Marketing. YEAR 2009 REC. 5 YDS. 42 AVG. 8.4 20 TIGHT END Peru, Ind. (Rochester) TD 1 ZACH WATKINS 6-2, 205, Jr. REC. 28 78 106 YDS. 329 918 1247 AVG. 11.8 11.8 11.8 RYAN WEBB 6-3, 235, So. TD 0 10 10 DEFENSIVE END Batavia, Ill. (Batavia) 2009: Played in 10 games as a tight end...had one reception for two yards at Albion...Personal: Marketing major...born 8/31/90...son of Ken and Dawn Webb...High School: AllConference. 14 YDS. 2378 AVG. 37.7 LONG 69 Additional Returning Players 63 NICK ATKINSON OFFENSIVE LINE 6-2, 270, R-So. Kokomo, Ind. (Western) 2009: Played in Butler’s first four games...named to the PFL Academic Honor Roll...2008: Did not see varsity action during his initial season with the Bulldogs... Personal: Psychology major...born 8/25/89...son of Scott Atkinson...High School: Three-year . 40 MATT BENSON RUNNING BACK 6-2, 215, R-Fr. Traverse City, Mich. (St. Francis) 2009: Did not see varsity action during his initial season with the Bulldogs...Personal: Business major...born 12/13/90...son of Charles and Barbara Benson... High School: All-Conference tight end at Travers City St. Francis High School... has 22 receptions for 501 yards and 10 touchdowns...earned nine varsity letters in three sports...three-year team captain and All-State performer on high school rugby team...High School Coaches: Josh Sellers/Greg Vaughan. 18 88 PUNTS 63 WIDE RECEIVER Chicago, Ill. (St. Ignatius) 2009 Career PUNTER Grand Rapids, Mich. (East Grand Rapids) 2009: Butler’s Special Teams Most Valuable Player...honorable mention All-PFL punter...ranked seventh in the PFL in punting average...led the league in number of punts inside the 20-yard line (29)...dropped four punts inside the 20-yard line in Butler’s victory at Dayton...had seven punts of 50 or more yards with a long punt of 69 yards against Davidson...named to the PFL Academic Honor Roll..to-back state championships in 2005 and 2006...named to the 2007 Dream Team... set a Michigan state high school record for most PATs (179)...averaged 41.0 yards as a punter in 2007...honor student...High School Coach: Peter Stuursma. YEAR 2009 11 MICHAEL WILSON 6-0, 180, So. CALVIN BLAIR QUARTERBACK 6-2, 210, R-So. Grand Rapids, Mich. (East Grand Rapids) 2009: Saw action in Butler’s first four games...completed 12 of 17 passes (.706) for 77 yards...threw an eight-yard touchdown pass in Butler’s victory over Hanover and a 10-yard scoring pass in the Bulldogs’ win at Franklin...named to the PFL Academic Honor Roll...2008: Did not see varsity action for the Bulldogs...named to the Pioneer Football League Academic Honor Roll...Personal: Business exploratory. 2010 Bulldogs 21 CHRIS BURNS DEFENSIVE BACK 5-9, 190, So. Aurora, Ill. (Waubonsee Valley). 69 JOSH DORFMAN OFFENSIVE LINE 6-0, 260, Jr. Durham, N.C. (C. E. Jordan) 2009: Played in Butler’s first three games...named to the PFL Academic Honor Roll...2008: Did not see varsity action...named to the PFL Academic Honor Roll...2007: Did not see varsity action in his initial season with the Bulldogs... Personal: Finance. 27 SEAN GRADY DEFENSIVE BACK 5-10, 180, R-Fr. Geneva, Ill. (Geneva) 2009: Did not see varsity action in first season with the Bulldogs...named to the PFL Academic Honor Roll...Personal: Education major...born 8/10/91...son of Brian and Maureen Grady...High School:. 29 DAN HABER DEFENSIVE BACK 6-0, 175, R-Fr. Calabasas, Calif. (Chaminade Prep) 2009: Chosen Butler’s Special Teams “Unsung Hero” (Scout Team Player of the Year)...did not see varsity action in first season with the Bulldogs...named to the PFL Academic Honor Roll...Personal: Business major...born 4/2/91...son of Rosemary Haber and Mitchell Haber...High School: Four-year Scholar-Athlete Award winner... recorded 32 tackles and eight pass breakups as a senior...earned varsity letters in football, basketball, baseball and volleyball...baseball team captain...High School Coach: Anthony Harris. 91 TAYLOR HARRIS DEFENSIVE TACKLE T 6-5, 290, Jr. Logansport, Ind. (Logansport) 2009: Did not see varsity action... Captains Corner Offensive guard Donnie Gilmore and defensive back Tadd Dombart were named captains for the 2010 Bulldogs, by a vote of their teammates. Gilmore, a fifth-year senior, has started every Butler game for the past three seasons, while Dombart has played in 33 games over the past three years, including 24 as a starter. 17 SCOTT HARVEY WIDE RECEIVER 6-3, 190, R-Fr. Carmel, Ind. (Westfield) 2009: Did not see varsity action in his initial season at Butler...named to the PFL Academic Honor Roll...Personal: Finance major...born 10/8/90...son of Bart and Jayne Harvey...High School: Starting quarterback. 3 TOM JUDGE QUARTERBACK 6-2, 190, R-Fr. Elmhurst, Ill. (York) 2009: Did not see varsity action in his initial season at Butler...Personal: Undecided. 13 T. J. LUKASIK DEFENSIVE BACK 5-8, 180, So. Lowell, Ind. (Lowell) 2009: Did not see varsity action...named to the PFL Academic Honor Roll... 30 BOBBY McDONALD DEFENSIVE BACK 6-0, 190, R-Fr. Orland Park, Ill. (Providence Catholic). 79 RYAN MYERS OFFENSIVE LINE 6-2, 285, So. Fort Wayne, Ind. (Bishop Dwenger) 2009: Saw action in Butler’s first three games...2008: Did not see varsity action... Personal:. 15 2010 Bulldogs 43 ALEX PERRITT LINEBACKER 6-1, 215, R-Fr. Whiteland, Ind. (Whiteland) 2009: Did not see varsity action in his first season at Butler...Personal: Biology major...born 2/8/91...son of Dan and Robin Perritt...High School: Honorable mention All-State performer at Whiteland. 75 DOUG PETTY OFFENSIVE LINE 6-4, 220, R-Fr. Indianapolis, Ind. (Southport) 2009: Did not see varsity action in his initial season at Butler...Personal: Biology/PreMed. 24 ANDREW PRATT DEFENSIVE BACK 5-10, 180, Jr. Ridge Farm, Ill. (Georgetown-Ridge Farm) 2009: Played in Butler’s first two games...credited with a pass breakup against Albion...2008: Did not see varsity action...2007: Did not see varsity action... named a 2007 Josten’s Scholar-Athlete...Personal: Business/Finance. 26 MIKE ROSE DEFENSIVE BACK 6-2, 200, R-Fr. Canton, Mich. (Plymouth) 2009: Did not see varsity action in his first season at Butler...named to the PFL Academic Honor Roll...Personal: Business major...born 11/28/90...son of Francine Rose...High School: All-Conference safety at Plymouth High School...helped high school squad to a Division championship as a senior...earned All-Conference honors on high school baseball squad...High School Coach: Mike Sawchuk. 9 LUCAS RUSKE QUARTERBACK 6-1, 190, R-Fr. Wilmette, Ill. (Loyola Academy) 2009: Did not see varsity action...Personal: Undecided major...born 7/15/91... son of Gary and Elizabeth Ruske...High School: more than 1,000 yards...had 13 rushing touchdown and 15 passing TDs...Academic All-State and Scholar-Athlete award recipient...High School Coach: John Holecek. 77 NICK SCHIRMANN OFFENSIVE LINE 6-1, 275, So. Cincinnati, Ohio (Anderson) 2009: Played in five of Butler’s first six games...Personal: Business major...born 3/26/91...son of Mike and Patty Schirmann...High School: All-State offensive lineman...helped lead Anderson to a 13-2 record and a state championship in 2007 and a 12-3 mark and a state runner-up finish in 2008...two-year team captain...twoyear All-Conference, All-City and All-District performer...received Anthony Munoz Foundation DII Offensive Lineman of the Year Award as a senior...earned three varsity football letters...High School Coach: Jeff Giesting. 16 87 CHARLIE SCHMELZER TIGHT END 6-7, 255, So. Bloomington, Ill. (Bloomington) 2009: Played in Butler’s first three games and in four total contests...named to the PFL Academic Honor Roll..Athlete...earned two varsity letters each in football and basketball...High School Coach: Rigo Schmelzer. 65 PAUL SCIORTINO OFFENSIVE LINE 6-1, 265, R-Fr. Wilmette, Ill. (Loyola Academy) 2009: Did not see varsity action in his first season with the Bulldogs...named to the PFL Academic Honor Roll...Personal: Business major...born 6/30/91...son of Bill and Celeste Sciortino...High School: in baseball...earned two varsity letters each in football and baseball...High School Coach: John Holecek. 84 BRENDAN SHANNON WIDE RECEIVER 5-11, 190, R-Fr. Lombard, Ill. (Montini Catholic) 2009: Did not see varsity action during his initial season with the Bulldogs...named to the PFL Academic Honor Roll...Personal: Undecided. 73 MATT STOREY OFFENSIVE LINE 6-3, 275, So. St. Louis, Mo. (St. Louis University H.S.). 23 DAVID THOMAS RUNNING BACK 5-9, 185, R-Fr. Chicago, Ill. (North Shore Country Day). 2010 Bulldogs 39 BRETT THOMASTON PLACEKICKER/PUNTER 6-5, 205, So. Frankfort, Ill. (Lincoln-Way East) 2009: Played in five of Butler’s first six contests, handling kickoffs...Personal: Pre-Physical Therapy. 82 JEFF URCH TIGHT END 6-4, 250, R-Fr. Avon, Ind. (Avon) 2009: Did not see varsity action in his initial season with the Bulldogs...Personal: Undecided. Newcomers GREG AMBROSE, OL, 6-3, 275, Fr., Dayton, Ohio (Oakwood)...Three-year starting offensive lineman at Oakwood High School...named conference Lineman of the Year...first team All-Southwest Ohio...HM All-State...team captain as a senior... High School Coach: Paul Stone. BRYCE BERRY BERRY, DB, 6-2, 190, Fr., St. Charles, Ill. (St. Charles East)...AllConference. JAY BRUMMEL, DB, 6-1, 210, R-Fr., Cedar Falls, Iowa (Cedar Falls/Iowa State)... Transfer from Iowa State where he sat out initial collegiate season as a redshirt... started one season at defensive back and one season at running back at Cedar Falls High School...2008 second team INA All-State selection...helped high school squad to a 12-2 record and a 4A state runner-up finish as a senior...played on 111 squad as a junior...four-year track and three-year wrestling letterwinner...High School Coaches: Pat Mitchell and Brad Remmert. DAVID BURKE, DB, 6-2, 180, Fr., Cincinnati, Ohio (Lakota West)...Two-year varsity starter at Lakota West High School...named Most Improved Varsity Defensive Player as a senior...second team All-Conference performer...helped high school squad to its first conference championship and a school-record nine wins in 2009... All-Academic Award winner...High School Coach: Larry Cox. BILL BUSCH, LB, 6-0, 238, Fr., St. Louis, Mo. (St. Louis Priory)...All-State linebacker at St. Louis Priory...named first team All-ABC League on defense...led high school squad in tackles in junior and senior seasons...helped lead St. Louis Priory to a 9-4 record and the state semifinals in 2009...played on high school teams that compiled a 26-10 record over three seasons...High School Coach: Marty Combs. JOHN CANNOVA, OL, 6-0, 260, Fr., Wheaton, Ill. (Benet Academy)...AllConference and All-Area offensive lineman at Benet Academy...team captain as a senior...three-year letterwinner on high school track and field squad...High School Coach: Gary Goforth. JOSEPH CIANCIO, DB, 5-10, 190, Fr., Darien, Ill. (Downers Grove South).... CALEB CONWAY CONWAY, LB, 5-9, 190, Fr., River Forest, Ill. (Oak Park-River Forest)... Two-year letterwinner as a linebacker at Oak Park-River Forest High School... team captain as a senior...finished second on the team in tackles (69) as a senior...tied for the team lead in quarterback sacks (8) in 2009...earned Academic All-Conference recognition...sprinter on high school track squad...High School Coach: Jim Nudera. KEVIN COOK, DB, 6-1, 180, Fr., Fishers, Ind. (Hamilton Southeastern)...All-State defensive back at Hamilton Southeastern High School...three-time All-Conference and two-time All-County performer...team captain as a senior...recorded 11 career pass interceptions...had a school-record 100-yard kickoff return for a touchdown... member of high school’s lacrosse and baseball squads...High School Coach: Scott May. CHRISTIAN EBLE, LB, 5-11, 220, Fr., Naperville, Ill. (Neuqua Valley)...Most Valuable Defensive Lineman as a nose guard at Neuqua Valley High School in 2009...HM All-State...first team All-Conference and All-City as a senior...HM AllArea and All-Region...named a “Top 50” player by both the Chicago Tribune and the Chicago Daily Herald...recorded 50 tackles as a senior, including 22 tackles for loss and five quarterback sacks...earned Academic All-Conference honors...High School Coach: Bryan Wells. GREG EGAN, LB, 5-11, 180, Fr., Naperville, Ill. (Neuqua Valley)...Two-year starting safety at Neuqua Valley High School...team captain and HM All-Conference performer...recorded 54 tackles and scored one defensive touchdown as a senior... helped high school track squad to a state runner-up finish in 2009...High School Coach: Bryan Wells. BRANDON GRUBBE, RB, 5-11, 195, Fr., Lowell, Ind. (Lowell)...All-time leading rusher at Lowell High School...first team All-State running back in 2009...named to the Indiana Football Coaches Association (IFCA) “Top 50” as a senior...three-time first team All-Conference and first team All-Area performer...two-time Junior All-State pick...team captain and Most Valuable Player...helped lead high school squad to a 39-5 record, three conference championships and two state runner-up finishes in three seasons...received Top Scholar-Athlete Award in 2007...varsity letterwinner in basketball and baseball...High School Coach: Kirk Kennedy. TRAE HEETER, RB, 5-9, 185, Fr., Indianapolis, Ind. (Lawrence North)...All-State running back at Lawrence North High School...team captain, MVP and All-Conference as a senior...Junior All-State and HM All-Conference in 2008...rushed for 933 yards and nine TDs as a senior...scored 12 touchdowns as a junior...set a school record with a 99-yard kickoff return for a touchdown...Academic Scholar-Athlete...member of high school track and field squad...High School Coach: Tom Dilley. MATT HITTINGER, WR, 6-2, 180, Fr., Valparaiso, Ind. (Valparaiso)...Two-way starter as a wide receiver and defensive back at Valparaiso High School...threetime HM All-Area on both offense and defense...first team All-Conference...two-time HM All-State...IFCA Region One North-South All-Star...team captain, MVP and Mental Attitude Award winner...had 43 career receptions and seven career pass interceptions...helped team to a school record for victories (9) and a conference championship in 2008...High School Coach: Mark Hoffman. JAY HOWARD, OL, 6-4, 245, Fr., Dayton, Ohio (Oakwood)... 17 2010 Bulldogs KYLE JACHIM, WR, 6-0, 175, Fr., Chicago, Ill. (St. Rita)... DYLAN JOHNSON, TE, 6-3, 215, So., Bloomington, Ill. (Central Catholic/ Lindenwood)...Transferred to Butler from Lindenwood University... WADE MARKLEY, QB, 6-4, 210, Fr., Fort Wayne, Ind. (Dwenger)... JT MESCH, WR, 6-2, 185, Fr., Glen Ellyn, Ill. (Glenbard West)... ARTHUR MONACO, RB, 5-9, 175, Fr., Bloomingdale, Ill. (Lake Park).. touchdown...helped Lake Park to a 7-1 conference record and an 8-3 overall mark in 2009...All-Conference performer on high school baseball squad...High School Coach: Andy Livingston. NICK NYKAZA, OL, 6-0, 235, Fr., Wilmette, Ill. (New Trier)...Three-year starting lineman at New Trier High School...named to the All-Central Suburban South Team as an offensive lineman...chosen as an IHSA Scholar Athlete...played on high school’s lacrosse team...High School Coach: Matt Irvin. DEREK O’CONNOR, WR, 6-0, 180, Fr., Bourbannais, Ill. (Bishop McNamara).... CHARLES PERRECONE, OL, 6-2, 285, Fr., Roselle, Ill. (Lake Park)... LOGAN PERRY, LB, 5-11, 205, Fr., Columbus, Ind. (Columbus East).....had unbeaten regular seasons in 2008 and 2009...Academic Letterman...High School Coach: Bob Gaddis. 18 PHILLIP POWELL, LB, 6-2, 210, Fr., Indianapolis, Ind. (Lawrence Central)... All-County linebacker at Lawrence Central High School...team captain and second team All-Conference pick...named Lawrence Central Male Athlete of the Year in in 2009-10...recorded 203 tackles and 10 quarterback sacks...three-year varsity starter...two-time All-County baseball player...High School Coach: Jayson West. JOEY PURZE, WR, 6-0, 195, Fr., Clayton, Mo. (Clayton)...Two-year starter as a wide receiver at Clayton High School...two-time HM All-Conference...participated in the U.S. Army Combine...compiled 545 receiving yards in three seasons...member of high school’s golf team...High School Coach: Sam Horell. WILL SCHIERHOLZ, QB, 6-0, 215, Fr., St. Louis, Mo. (MICDS)... JIMMY SCHWABE, DB, 5-11, 190, Fr., Downers Grove, Ill. (Downers Grove South)...Two-year starter at Downers Grove South High School..missed all but three games of his senior season with injury...two-time team captain...helped high school squad to a pair of unbeaten conference championships...Academic All-Conference performer...High School Coach: John Belskis. JEREMY STEPHENS, DL, 6-1, 267, So., Indianapolis, Ind. (Lawrence Central)... Transfer from Thomas More College...played in nine games as a freshman at Thomas More and helped the team to an 11-1 season...recorded 22 tackles, including two quarterback sacks and four tackles for loss, in first collegiate season...earned three varsity letters as a defensive lineman at Lawrence Central High School... High School Coach: Jayson West. DON STEWART, DB, 5-10, 175, Fr., St. Louis, Mo. (Clayton)... JAYME SZAFRANSKI, DB, 6-1, 190, Fr., Barrington, Ill. (Fremd)... MIKE WENDAHL, OL, 6-5, 290, Jr., Valparaiso, Ind. (Valparaiso/Indianapolis)... Transfer from the University of Indianapolis...will sit out the 2010 season...two-time All-State offensive lineman at Valparaiso High School...named All-Conference and All-Area in junior and senior seasons...team captain in final prep season...chosen to play in the Indiana North-South All-Star game...earned three varsity letters in football... member of high school basketball team...High School Coach: Mark Hoffman. DANIEL WILSON, PK, 5-9, 198, Fr., Zionsville, Ind. (Zionsville)...Two-year letterwinner as a placekicker at Zionsville High School...tied a school record with a 44-yard field goal...earned Academic All-State recognition...High School Coach: Larry McWhorter. PAUL YANOW, LB, 6-0, 200, Fr., Cincinnati, Ohio (Sycamore)...Two-year ...Two-year leading ...T tackler at Sycamore High School...named high school team’s Interior Defensive Player of the Year in back-to-back seasons...two-year second team All-Conference...HM All-City as a senior...Sycamore Old Spice Player of the Year in 2009...chosen to play in the Ohio East-West All-Star Game...four-time Academic All-Conference honoree... two-year varsity letterwinner in baseball...High School Coach: Scott Datillo. 2009 Gridiron Classic Butler 28, Central Connecticut State 23 Safety Mark Giacomantonio (above) led Butler’s defensive effort against Central Connecticut State with a team-high 10 tackles. Runningback Scott Gray (above) rushed for a teamhigh 88 yards and scored the Bulldogs’ first touchdown in the Gridiron Classic. Quarterback Andrew Huck (above) scored the final of his three touchdowns to clinch Butler’s first postseason football victory. Quarterback Andrew Huck (above) had over 200 yards in total offense and was named the Most Valuable Player of the 2009 Gridiron Classic. Wide receiver Dan Bohrer (above) had a team-high seven receptions for 69 yards and defensive end Grant Hunter (left) added five tackles against CCSU. Linebacker Nick Caldicott (above) had eight tackles in Butler’s winning effort in the 2009 Gridiron Classic. 19 2010 Butler Roster Numerical Roster/ Pronunciation Guide 1 2 3 4 5 6-D 6-O 7-D 7-O 8 9-D 9-O 10 11 12 13 14 15-D 15-O 16 17-D 17-O 18 19 20 21-D 21-O 22 23-D 23-O 24 25 26 27 28 29 30 31-D 31-O 32-D 32-O 33 34-D 34-O 35 36 37 38 39 40 41 42 43 44 20 Jordan Koopman Stuart Harvey Tom Judge Matt Foor Artis Hailey III Logan Sullivan Wade Markley Jimmy Schwabe (Shwab-ee) Jeff Larsen Ryan Hitchcock Andy Dauch (DOWK) Lucas Ruske (Russ-KEE) Andrew Huck Zach Watkins Matt Kobli (KOE-blee) T. J. Lukasik (Loo-CAY-sik) Nick Caldicott (KAL-dah-cott) Andrew Cottrell Will Schierholz (SHEER-holz) Mark Giacomantonio (JOK-ah-mon-TOE-nee-oh) Bryce Berry Scott Harvey Calvin Blair Joseph Ciancio (See-AHN-SEE-oh) Michael Wilson Chris Burns Joseph Purze (PURRS) Scott Gray David Burke David Thomas Andrew Pratt Jack McKenna Mike Rose Sean Grady Tadd Dombart Dan Haber Bobby McDonald Caleb Conway Brandon Grubbe (GROO-bee) Don Stewart Trae Heeter William Bork Jay Brummel Arthur Monaco David Lang Steven Depositar Jayme Szafranski (ZAH-fran-skee) Kevin Cook Brett Thomaston Matt Benson Bill Busch Phillip Powell Alex Perritt Greg Egan NO. 66 63 17-D 40 18 NAME Greg Ambrose Nick Atkinson Bryce Barry Matt Benson Calvin Blair POS. OL OL DB RB QB HT. 6-3 6-2 6-2 6-2 6-2 WT. 275 280 190 220 205 YR. Fr. So. Fr. R-Fr. So. HOMETOWN/HIGH SCHOOL Dayton, Ohio/Oakwood Kokomo, Ind./Western St. Charles, Ill./St. Charles East Traverse City, Mich./St. Francis Grand Rapids, Mich./East Grand Rapids 33 34-D 23-D 21-D 41 William Bork** Jay Brummel David Burke Chris Burns Bill Busch TE DB DB DB LB 6-1 6-1 6-2 5-9 6-0 225 210 180 180 238 Jr. R-Fr. Fr. So. Fr. Dyer, Ind./Lake Central Cedar Falls, Iowa/Cedar Falls/Iowa State Cincinnati, Ohio/Lakota West Aurora, Ill./Waubonsee Valley St. Louis, Mo./St. Louis Priory 14 71 19 95 31-D Nick Caldicott** John Cannova Joseph Ciancio Taylor Clarkson* Caleb Conway LB OL DB DT LB 6-1 6-0 5-10 6-2 5-9 225 260 190 275 190 Jr. Fr. Fr. So. Fr. Oak Park, Ill./Fenwick Wheaton, Ill./Benet Academy Darien, Ill./Downers Grove South Fortville, Ind./Mt. Vernon River Forest, Ill./Oak Park-River Forest 38 46 15-D 9-D 36 Kevin Cook Rob Cosler** Andrew Cottrell* Andy Dauch** Steven Depositar* DB DE LB CB RB 6-1 6-1 5-11 5-10 5-10 180 230 215 180 195 Fr. Sr. Jr. Jr. So. Fishers, Ind./Hamilton Southeastern Winter Park, Fla./Winter Park West Chester, Ohio/Lakota W./Grand Valley Bloomfield Hills, Mich./Lahser Mishawaka, Ind./Penn 28 69 49 44 4 Tadd Dombart*** Josh Dorfman Christian Eble Greg Egan Matthew Foor* CB OL LB LB DB 5-10 6-0 5-11 5-11 5-11 185 260 220 180 205 Sr. Jr. Fr. Fr. So. Cincinnati, Ohio/Lakota West Durham, N.C./C.E. Jordan Naperville, Ill./Neuqua Valley Naperville, Ill./Neuqua Valley West Chester, Ohio/Lakota West 16 64 27 22 31-O Mark Giacomantonio** Donnie Gilmore*** Sean Grady Scott Gray*** Brandon Grubbe SS OL CB RB RB 5-11 6-3 5-10 5-8 5-11 190 295 190 185 195 Jr. Sr. R-Fr. Sr. Fr. Carmel, Ind./Carmel Westphalia, Ind./North Knox Geneva, Ill./Geneva Lowell, Ind./Lowell/Ball State Lowell, Ind./Lowell 29 5 91 17-O 2 Dan Haber Artis Hailey III* Taylor Harris Scott Harvey Stuart Harvey* CB LB DT WR WR 6-0 6-0 6-5 6-3 6-1 180 215 285 185 190 R-Fr. So. Jr. R-Fr. Jr. Calabasas, Calif./Chaminade Prep Hammond, Ind./Hammond Logansport, Ind./Logansport Carmel, Ind./Westfield Carmel, Ind./Westfield 32-O 8 86 53 67 Trae Heeter Ryan Hitchcock** Matt Hittinger Rob Hobson* Jay Howard RB RB WR OL OL 5-9 5-7 6-2 6-4 6-4 185 200 180 280 245 Fr. Jr. Fr. Sr. Fr. Indianapolis, Ind./Lawrence North Bedford, Ind./Bedford-North Lawrence Valparaiso, Ind./Valparaiso Arcadia, Ind./Sheridan Dayton, Ohio/Oakwood 10 98 80 93 3 Andrew Huck* Grant Hunter** Kyle Jachim Dylan Johnson# Tom Judge QB DE WR TE QB 6-2 6-3 6-0 6-3 6-2 195 240 175 215 190 Jr. Jr. Fr. So. R-Fr. Bloomington, Ind./Bloomington North Liberty Twp., Ohio/Lakota West/Robert Morris Chicago, Ill./St. Rita Bloomington, Ill./Cent. Catholic/Lindenwood Elmhurst, Ill./York 12 1 51 54 35 Matt Kobli** Jordan Koopman** Robert Koteff* William Lamar* David Lang** QB WR LB LB PK/P 6-3 5-8 6-0 5-11 6-0 240 190 225 200 220 Sr. Jr. Jr. So. Jr. Whiting, Ind./Whiting Chula Vista, Calif./Eastlake Lexington, Ky./Paul L. Dunbar Oak Park, Ill./Oak Park-River Forest Lowell, Ind./Lowell 7-O 13 6-O 72 30 Jeff Larsen* T. J. Lukasik Wade Markley Pete Mattingly** Bobby McDonald WR DB QB OL DB 6-1 5-8 6-4 6-5 6-0 190 180 210 290 200 Jr. So. Fr. Jr. R-Fr. Chicago, Ill./Notre Dame Lowell, Ind./Lowell Fort Wayne, Ind./Dwenger Zionsville, Ind./Cathedral Orland Park, Ill./Providence Catholic 2010 Butler Roster NO. 81 25 89 34-O 79 NAME Eddie McHale*** Jack McKenna** JT Mesch Arthur Monaco Ryan Myers POS. WR CB WR RB OL HT. 6-3 6-0 6-2 5-9 6-2 WT. 205 180 185 175 290 YR. Sr. R-Jr. Fr. Fr. So. HOMETOWN/HIGH SCHOOL Cincinnati, Ohio/St. Xavier/Ohio Elmwood Park, Ill./St. Patrick Glen Ellyn, Ill./Glenbard West Bloomingdale, Ill./Lake Park Fort Wayne, Ind./Bishop Dwenger 68 83 47 62 43 Nick Nykaza Derek O’Connor Bob Olszewski* Charles Perrecone Alex Perritt OL WR LB OL LB 6-0 6-0 6-1 6-2 6-1 235 180 205 285 215 Fr. Fr. Sr. Fr. R-Fr. Wilmette, Ill./New Trier Bourbannais, Ill./Bishop McNamara LaGrange, Ill./Lyons Township Roselle, Ill./Lake Park Whiteland, Ind./Whiteland 48 75 52 42 24 Logan Perry Doug Petty Jeff Poss** Phillip Powell Andrew Pratt LB OL DE LB FS 5-11 6-4 6-2 6-2 5-10 205 240 240 210 180 Fr. R-Fr. Jr. Fr. Jr. Columbus, Ind./Columbus East Indianapolis, Ind./Southport Salt Lake City, Utah/Scripps Ranch (Calif.) Indianapolis, Ind./Lawrence Central Ridge Farm, Ill./Georgetown-Ridge Farm 21-O 59 26 9-O 15-O Joseph Purze Jordan Ridley* Mike Rose Lucas Ruske Will Schierholz WR LB FS QB QB 6-0 5-11 6-2 6-1 6-0 195 230 200 190 215 Fr. So. R-Fr. R-Fr. Fr. Clayton, Mo./Clayton Indianapolis, Ind./Lawrence Central Canton, Mich./Plymouth Wilmette, Ill./Loyola Academy St. Louis, Mo./MICDS 77 87 7-D 65 76 Nick Schirmann Charlie Schmelzer Jimmy Schwabe Paul Sciortino Ryan Secrist** OL TE DB OL OL 6-1 6-7 5-11 6-1 6-4 275 255 190 265 275 So. So. Fr. R-Fr. Sr. Cincinnati, Ohio/Anderson Bloomington, Ill./Bloomington Downers Grove, Ill./Downers Grove South Wilmette, Ill./Loyola Academy Rochester, Ind./Rochester 84 97 78 56 32-D Brendan Shannon Tyler Skaggs* Mike Staniewicz*** Jeremy Stephens# Don Stewart WR DT OL DL DB 5-11 6-2 6-6 6-1 5-10 190 240 285 267 175 R-Fr. Sr. Sr. So. Fr. Lombard, Ill./Montini Catholic Winona Lake, Ind./Warsaw Hebron, Ind./Lowell Indianapolis, Ind./Lawrence C./Thomas More St. Louis, Mo./Clayton 73 6-D 37 57 55 Matt Storey Logan Sullivan** Jayme Szafranski Ross Teare** Jace Tennant* OL SS DB DT DE 6-3 6-0 6-1 6-1 6-1 275 195 190 290 225 So. Jr. Fr. Jr. So. St.Louis, Mo./St. Louis University H.S. Winston-Salem, N.C./Horizon (Ariz.) Barrington, Ill./Fremd Sammamish, Wash./Eastlake Champaign, Ill./Central 23-O 90 39 82 94 David Thomas Larry Thomas Brett Thomaston Jeff Urch Carter Walley* RB DT K/P TE TE 5-9 5-10 6-5 6-4 6-2 185 250 205 250 255 R-Fr. Jr. So. R-Fr. So. Chicago, Ill./North Shore Country Day West Chester, Ohio/St. Xavier Frankfort, Ill./Lincoln-Way East Avon, Ind./Avon Peru, Ind./Rochester 11 88-D 61 96 20 45 Zach Watkins** Ryan Webb* Mike Wendahl# Daniel Wilson Michael Wilson* Paul Yanow WR DE OL PK P LB 6-2 6-3 6-5 5-9 6-0 6-0 205 235 290 198 180 200 Jr. So. So. Fr. So. Fr. Chicago, Ill./St. Ignatius Batavia, Ill./Batavia Valparaiso, Ind./Valparaiso/Indianapolis Zionsville, Ind./Zionsville Grand Rapids, Mich./East Grand Rapids Cincinnati, Ohio/Sycamore 45 46 47 48 49 51 52 53 54 55 56 57 59 61 62 63 64 65 66 67 68 69 71 72 73 75 76 77 78 79 80 81 82 83 84 86 87 88 89 90 91 93 94 95 96 97 98 Paul Yanow (YEAH-now) Rob Cosler Bob Olszewski (OLE-chev-skee) Logan Perry Christian Eble (EHBLE) Robert Koteff (COE-teff) Jeff Poss Rob Hobson William Lamar Jace Tennant Jeremy Stephens Ross Teare (TEER) Jordan Ridley Mike Wendahl Charles Perrecone (Purr-CONE-ee) Nick Atkinson Donnie Gilmore Paul Sciortino (SCORE-in-teen-oh) Greg Ambrose Jay Howard Nick Nykaza (NYE-kah-zah) Josh Dorfman John Cannova (Cah-NO-vah) Pete Mattingly Matt Storey Doug Petty Ryan Secrist Nick Schirmann Mike Staniewicz (STAN-ah-witts) Ryan Myers Kyle Jachim (YAH-kum) Eddie McHale Jeff Urch Derek O’Connor Brendan Shannon Matt Hittinger (HIT-in-jer) Charlie Schmelzer Ryan Webb JT Mesch Larry Thomas Taylor Harris Dylan Johnson Carter Walley (WALLY) Taylor Clarkson Daniel Wilson Tyler Skaggs Grant Hunter *—Varsity letters earned. #--Not eligible until 2011. O--Number on offense. D--Number on defense. HEAD COACH: Jeff Voris (DePauw ’89) ASSISTANT COACHES: Joe Cheshire, Tim Cooper, Nick Tabacca, David Kenney, Nick Anderson, Danny Sears, Chris Davis, Rob Noel, Matt Walker 21 2009 Butler Statistics SCORING Zach Watkins Andrew Huck David Lang Dan Bohrer Scott Gray Ryan Hitchcock Eddie McHale Ricky Trujillo Jordan Koopman Carter Walley Andy Dauch Logan Sullivan Michael Wilson Butler Opponents FGs David Lang Butler Opponents G TD 12 10 12 10 12 0 12 7 12 6 10 4 12 3 11 2 12 2 12 1 12 1 12 0 12 0 12 46 12 30 PAT-1 0-0 0-0 33-43 0-0 0-0 0-0 0-0 0-0 0-0 0-0 0-0 0-0 0-1 33-44 24-24 PAT-2 0-0 0-0 0-0 0-0 0-0 0-0 0-0 0-0 0-0 0-0 0-0 1-2 0-0 1-2 2-6 FG PTS 0-0 60 0-0 60 7-11 54 0-0 42 0-0 36 0-0 24 0-0 18 0-0 12 0-0 12 0-0 6 0-0 6 0-0 2 0-0 0 7-11 332 7-11 229 G 10-19 20-29 30-39 40-49 50+ LG 12 0-0 4-4 3-6 0-1 0-0 39 12 0-0 4-4 3-6 0-1 0-0 39 12 0-0 2-2 3-4 2-4 0-1 44 RUSHING Scott Gray Andrew Huck Ryan Hitchcock Steven Depositar Ricky Trujillo Dan Bohrer Artis Hailey Jordan Koopman Calvin Blair Jordan Ridley Team Butler Opponents G 12 12 10 12 11 12 12 12 4 12 12 12 12 ATT GAIN LOSS NET TD AVG 169 913 45 868 5 5.1 102 580 113 467 10 4.6 55 350 4 346 4 6.3 19 132 2 130 0 6.8 24 127 2 125 2 5.2 13 90 0 90 0 6.9 13 65 0 65 0 5.0 5 62 5 57 1 11.4 6 21 4 17 0 5.7 1 13 0 13 0 13.0 7 0 65 -65 0 --414 2353 240 2113 22 5.1 425 1709 310 1399 13 3.3 PASSING Andrew Huck Calvin Blair Team Butler Opponents G 12 4 12 12 12 ATT COMP PCT. 371 233 .628 17 12 .706 2 0 .000 390 245 .628 365 196 .537 INT 11 0 0 11 18 YDS. 2454 77 0 2531 2265 RECEIVING Zach Watkins Dan Bohrer Eddie McHale Jordan Koopman Scott Gray Jeff Larsen Ryan Hitchcock Ricky Trujillo Carter Walley Artis Hailey William Bork Ryan Webb Butler Opponents G 12 12 12 12 12 12 10 11 12 12 12 10 12 12 NO. 78 52 38 24 13 13 9 7 5 3 2 1 245 196 PUNTING Michael WIlson Team Butler Opponents G 12 12 12 12 NO. 63 1 64 74 YDS. 2378 0 2378 2637 AVG. 37.7 0.0 37.2 35.6 LG 69 0 69 63 PUNT RETURNS Tadd Dombart Jordan Koopman Logan Sullivan Andy Dauch Team Butler Opponents G 12 12 12 12 12 12 12 NO. 11 8 1 1 2 23 24 YDS. 66 104 15 11 0 196 147 AVG. 6.0 13.0 15.0 11.0 0.0 8.5 6.1 LG 16 34 0 11 0 34 25 22 YDS. TD 918 10 471 7 493 3 258 1 134 1 112 0 50 0 36 0 42 1 4 0 11 0 2 0 2531 23 2265 14 TD 21 2 0 23 14 AVG. LG 11.8 62 9.1 28 13.0 42 10.8 34 10.3 71 8.6 27 5.6 35 5.1 10 8.4 16 1.3 5 5.5 10 2.0 2 10.3 71 11.6 62 KICKOFF RETURNS Jordan Koopman Andy Dauch Zach Watkins Butler Opponents G 12 12 12 12 12 NO. 31 10 1 42 51 TEAM STATISTICS First Downs By Rush By Pass By Penalty Rush Yards Per Game Pass Yards Per Game Points Per Game Total Plays Total Offense Offense Per Play Offense Per Game Fumbles/Lost Penalties/Yards Third Down Conversions Fourth Down Conversions Time of Possession/Game PLAYER Nick Caldicott Derek Guggenberger Spencer Summerville Mark Giacomantonio Andrew Cottrell Jack McKenna Jeff Poss Grant Hunter Logan Sullivan Jordan Ridley Jacob Fritz Tadd Dombart Brian Adika Ross Teare Nick Comotto Rob Cosler William Lamar Pete Xander Larry Thomas Jace Tennant Matt Foor Thorn Murphy Derek Bradford Andy Dauch Dan Bohrer Jeff Larsen Tyler Skaggs Zach Watkins Taylor Clarkson Kevin Credille Jordan Koopman Brian Crable Pete Mattingly Bob Olszewski David Lang Andrew Huck Donnie Gilmore Eddie McHale Robert Koteff Ben Jones Brett Thomaston Andrew Pratt YDS. 703 174 0 877 998 BUTLER 255 111 118 26 176.1 210.9 27.7 804 4644 5.8 387.0 19/8 57/575 69/164 6/15 30:24 G 12 12 12 9 10 12 12 9 12 12 12 11 10 12 12 12 12 12 11 12 10 8 12 12 12 12 10 12 11 2 12 12 12 12 12 12 12 12 10 4 5 2 ST 39 31 36 30 16 31 24 19 22 20 18 20 10 9 12 15 10 7 8 10 7 6 6 5 3 2 1 3 2 2 0 1 1 1 1 1 0 0 0 0 0 0 AVG. 22.7 17.4 0.0 20.9 19.6 LG 62 28 0 62 46 OPPONENTS 202 78 110 14 116.6 188.8 19.1 790 3664 4.6 305.3 20/8 68/607 57/174 8/22 29:36 SCORE BY QUARTERS 1 2 Butler 47 108 Opponents 48 45 3 81 79 2009 RESULTS (11-1, 7-1 PFL) BU S. 5 - ALBION 42 S. 12 - At Franklin 49 S. 19 - HANOVER 42 S. 26 - At Morehead State* (OT) 28 O. 3 - SAN DIEGO* 25 O. 17 - VALPARAISO* 23 O. 24 - At Campbell* 23 O. 31 - DAVIDSON* 14 N. 7 - At Dayton* 31 N. 14 - At Jacksonville* 7 N. 21 - DRAKE* 20 D. 5 - CENTRAL CONNECTICUT ST.# 28 4 89 57 OT 7 0 OPP. 3 19 21 21 24 14 16 7 28 36 17 23 T 332 229 ATTD. 2,886 2,500 2,215 3,880 4,218 1,723 4,851 2,568 4,012 1,933 2,122 1,577 *--Pioneer Football League game. #--Gridiron Classic. BUTLER UNIVERSITY DEFENSIVE STATISTICS TACKLES QB FORCED FUMBLE PASS AT TOTAL FOR LOSS SACKS FUMBLE REC. INT. 31 70 8-22 1-7 3 1 3-25 38 69 8-25 4-19 0 1 0-0 32 68 0-0 0-0 0 0 1-22 15 45 4-11 0-0 1 1 5-45 27 43 2.5-4 0-0 0 0 0-0 11 42 1-1 0-0 0 0 3-30 17 41 7.5-19 3-13 0 0 0-0 20 39 11.5-52 10-42 1 0 0-0 16 38 1-1 0-0 0 0 1-10 16 36 8.5-35 4-22 1 1 1-15 16 34 0-0 0-0 0 1 0-0 10 30 2.5-4 0-0 0 0 2-9 19 29 0-0 0-0 0 0 0-0 20 29 4-14 2-8 0 0 0-0 16 28 0.5-0 0-0 0 0 0-0 11 26 1.5-3 1-2 0 0 0-0 14 24 1-2 0-0 0 0 0-0 11 18 4-15 1-9 0 1 0-0 9 17 5.5-18 1-2 1 0 0-0 6 16 5-19 5-19 0 0 0-0 6 13 0-0 0-0 0 1 0-0 6 12 0-0 0-0 0 0 1-8 6 12 0-0 0-0 0 0 0-0 5 10 0-0 0-0 1 0 1-19 3 6 0-0 0-0 0 0 0-0 3 5 0-0 0-0 0 0 0-0 3 4 0.5-1 0-0 0 0 0-0 0 3 0-0 0-0 0 0 0-0 0 2 0-0 0-0 0 0 0-0 0 2 0-0 0-0 0 0 0-0 2 0 0 0-0 0-0 0 0 0-0 PASS BRKUP 1 0 2 2 0 5 1 3 4 0 2 3 2 0 2 1 0 0 0 0 0 0 0 1 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 Blocked Kicks: Logan Sullivan (Punt vs. Albion), Spencer Summerville (Punt vs. Morehead St.), William Lamar (Punt vs. Morehead St.) 2009 Butler Results BUTLER 42, ALBION 3 Sept. 5 at Butler Bowl (A - 2,886) Score By Quarters 1 2 3 4 F Albion 0 0 3 0 3 Butler 0 28 14 0 42 SCORING SUMMARY BU - Eddie McHale, 10 pass from Andrew Huck (David Lang kick) BU - Scott Gray, 71 pass from Huck (Lang kick) BU - Zach Watkins, 48 pass from Huck (Lang kick) BU - Zach Watkins, 62 pass from Huck (Lang kick) AC - Mychal Galla, 29 field goal BU - Zach Watkins, 19 pass from Huck (Lang kick) BU - Andrew Huck, 3 run (Lang kick) RUSHING: BU - Huck, 8-68; Gray, 9-34; Depositar, 6-33; Hailey, 4-17; Hitchcock, 3-17; Blair, 4-17; Trujillo, 2-9; Bohrer, 1-5. AC - Orr, 13-36; Frisbey, 7-22; Harris, 3-minus 8; Fusee, 1-minus 10. PASSING: BU - Huck, 18-25-1 for 327 yards. AC - Harris, 10-25-1 for 59 yards; Fusee, 9-16-1 for 61 yards. RECEIVING: BU - Watkins, 7-183; McHale, 3-30; Gray, 2-74; Bohrer, 2-14; Trujillo, 2-9; Koopman, 1-16; Webb, 1-2; Larsen, 1-1. AC - McRobb, 8-52; Wunderlich, 5-36; Pointer, 2-13; Cruse, 2-12; Mayhoe, 1-6; Orr, 1-1. BUTLER 49, FRANKLIN 19 Sept. 12 at Red Faught Stadium (A - 2,500) Score By Quarters 1 2 3 4 F Butler 10 13 14 12 49 Franklin 0 7 6 6 19 SCORING SUMMARY BU - David Lang, 28 field goal BU - Zach Watkins, 3 pass from Andrew Huck (Lang kick) BU - Dan Bohrer, 9 pass from Huck (Kick failed) FC - Ryan Momberger, 1 pass from Kyle Ray (M. Magdalinos kick) BU - Eddie McHale, 8 pass from Huck (Lang kick) BU - Watkins, 7 pass from Huck (Lang kick) BU - Ricky Trujillo, 8 run (Lang kick) FC - Nick Purichia, 8 run (Pass failed) BU - Andrew Huck, 5 run (Kick failed) FC - A. Mellencamp, 62 pass from Purichia (Pass ffailed) BU - Jordan Koopman, 10 pass from Calvin Blair (Kick Failed) RUSHING: BU - Depositar, 5-50; Huck, 8-45; Gray, 7-27; Trujillo, 7-20; Hailey, 3-19; Bohrer, 1-10; Team, 1-minus 10. FC - Purichia, 6-21; Snellenbarger, 3-8; Cook, 1-4; Downs, 3-3; Heller, 3-3; Ellis, 1-minus 9; Ray, 7-minus 16; Team, 2-minus 31. PASSING: BU - Huck, 33-48-0 for 316 yards; Blair, 5-5 for 25 yards. FC - Ray, 19-36-2 for 261 yards; Purichia, 6-16-0 for 98 yards. RECEIVING: BU - Watkins, 8-91; Bohrer, 8-72; McHale, 7-62; Koopman, 6-77; Larsen, 3-19; Gray, 3-11; Hailey, 2-4; Trujillo, 1-5. FC - Deffner, 7-152; Mellencamp, 6-113; Momberger, 5-49; Walton, 3-20; Zmich, 2-21; Cook, 1-7; Downs, 1-minus 3. BUTLER 42, HANOVER 21 Sept. 19 at Butler Bowl (A - 2,215) Score By Quarters 1 2 3 4 F Hanover 7 0 7 7 21 Butler 14 21 7 0 42 SCORING SUMMARY BU - Andy Dauch, 18 pass interception return (David Lang kick) HC - S. Gibson, 8 run (P. Polochanin kick) BU - Andrew Huck, 8 run (Lang kick) BU - Ricky Trujillo, 12 run (Lang kick) BU - Dan Bohrer, 3 pass from Huck (Lang kick) BU - Scott Gray, 1 run (Lang kick) BU - Dan Bohrer, 8 pass from Calvin alvin Blair (Lang kick) HC - B. Smart, 15 pass from D. Passafiume (P. Polochanin kick) HC - C. Zeck, 5 pass from D. Seay (P. Polochanin kick) RUSHING: BU - Depositar, 8-47; Gray, 6-35; Huck, 4-31; Hailey, 6-29; Trujillo, 4-27; Blair, 2-4. HC - Seay, 7-19; Cook, 7-11; Zeck, 1-7; Gibson, 13-1; Juett, 2-1; Armstrong, 2-minus 7; Team, 1-minus 1. PASSING: BU - Huck, 13-15-0 for 134 yards; Blair, 6-11-0 for 50 yards. HC - Gibson, 11-15-0 for 92 yards; Seay, 1-2-0 for 5 yards; Passafiume, 1-1 for 15 yards; Armstrong, 1-1 for 6 yards; Polochanin, 1-1 for 14 yards; Stewart, 0-1. RECEIVING: BU - Watkins, 5-53; Bohrer, 5-45; Koopman, 5-36; McHale, 2-47; Gray, 1-3; Hailey, 1-0. HC - Passafiume, 6-27; Robinette, 2-33; Miller, 2-28; Smart, 2-22; Zeck, 2-16; Cason, 1-6. BUTLER 28, MOREHEAD STATE 21 (OT) Sept. 26 at Jayne Stadium (A - 3,880) Score By Quarters 1 2 3 4 OT F Butler 0 0 13 8 7 28 Morehead State 21 0 0 0 0 21 SCORING SUMMARY MS - D. Sawyer, 60 pass from E. Sawyer (R. Duzan kick) MS - D. Harkness, 63 pass interception return (Duzan kick) MS - D. Morgan, 28 pass from E. Sawyer (Duzan kick) BU - Zach Watkins, 9 pass from Andrew Huck ( David Lang kick) BU - Dan Bohrer, 2 pass from Huck (Kick blocked) BU - Zach Watkins, 7 pass from Huck (Logan Sullivan rush) BU - Eddie McHale, 22 pass from Huck (Lang kick) RUSHING: BU - Gray, 14-59; Bohrer, 1-18; Trujillo, 2-5; Hitchcock, 1-0; Huck, 5-minus 12, Team, 1-minus 39. MS - Pendleton, 11-57, E. Sawyer, 18-38; Morgan, 1-9; Smart, 1-6; Bodrick, 1-4; Cox, 3-1; Lewis, 1-minus 4; McDermott, 1-minus 6. PASSING: BU - Huck, 26-49-3 for 208 yards. MS - E. Sawyer, 9-20-1 for 149 yards; Lewis, 2-7-2 for 14 yards. RECEIVING: BU - Watkins, 13-130; Bohrer, 5-22; McHale, 3-27; Hitchcock, 3-2; Gray, 1-13; Larsen, 1-4. MS - McDermott, 3-23; Yoshimura, 2-15; Bodrick, 2-13; D. Sawyer, 1-60; Morgan, 1-28; Williams, 1-21; Pendleton, 1-3. BUTLER 25, SAN DIEGO 24 Oct. 3 at Butler Bowl (A - 4,218) Score By Quarters 1 2 3 4 F San Diego 3 14 0 7 24 Butler 0 9 6 10 25 SCORING SUMMARY SD - Mike Levine, 44 field goal BU - David Lang, 39 field goal SD - JT Rogan, 1 run (Levine kick) BU - Scott Gray, 12 run (Kick failed) SD - Patrick Doyle, 13 pass from Sam Scudellari (Levine kick) BU - Ryan Hitchcock, 43 run (Kick failed) SD - Matt Jelmini, 11 run (Levine kick) BU - Zach Watkins, 21 pass from Andrew Huck (Lang kick) BU - David Lang, 37 field goal (0:01) RUSHING: BU -Gray, 16-113; Hitchcock, 5-50; Huck, 11-35; Ridley, 1-13; Bohrer, 1-4. SD - Jelmini, 22-112; Rogan, 12-49; Scudellari, 8-49; Fontenberry, 1-minus 1. PASSING: BU - Huck, 25-38-1 for 260 yards. SD - Scudellari, 1728-0 for 175 yards. RECEIVING: BU -Watkins, 8-97; Bohrer, 8-79; Koopman, 4-57; McHale, 2-18; Walley, 1-8; Hitchcock, 1-1; Gray, 1-0. SD - Fiege, 5-30; Doyle, 4-65; McGough, 4-49; Brown, 2-15; Smith, 1-16; Jelmini, 1-0. BUTLER 23, VALPARAISO 14 Oct. 17 at Butler Bowl (A - 1,723) Score By Quarters 1 2 3 4 F Valparaiso 0 0 14 0 14 Butler 3 0 7 13 23 SCORING SUMMARY BU - David Lang, 38 field goal VU - Eli Crawford, 60 pass interception return (Andrzej Skiba kick) VU - Quinn Schafer, 1 run (Skiba kick) BU - Ryan Hitchcock, 4 run (Lang kick) BU - Scott Gray, 7 run (Lang kick) BU - Ryan Hitchcock, 2 run (Kick failed) RUSHING: BU - Gray, 22-99; Huck, 9-76; Hitchcock, 7-53; Trujillo, 2-11; Bohrer, 2-9; Team, 1-minus 1. VU - Wildermuth, 23-87; Lynn, 7-37; Wysocki, 5-7; Morelli, 5-6; Jones, 1-5; Schafer, 4-minus 15. PASSING: BU - Huck, 10-20-3 for 77 yards. VU - Schafer, 3-11-1 for 74 yards; Wysocki, 5-10-1 for 48 yards. RECEIVING: BU - McHale, 4-25; Watkins, 3-22; Hitchcock, 1-20; Walley, 1-6; Bohrer, 1-4. VU - Henton, 2-63; McCarty, 2-25; Myers, 1-15; Bennett, 1-14; Jones, 1-4; Wildermuth, 1-1. BUTLER 23, CAMPBELL 16 Oct. 24 at Barker-Lane Stadium (A - 4,851) Score By Quarters 1 2 3 4 F Butler 13 7 0 3 23 Campbell 0 10 3 3 16 SCORING SUMMARY BU - Scott Gray, 5 run (Kick Failed) BU - Zach Watkins, 22 pass from Andrew Huck (David Lang kick) BU - Dan Bohrer, 1 pass from Huck (Lang kick) CU - CJ Oates, 9 pass from Daniel Polk (Adam Willets kick) CU - Adam WIllets, 38 field goal CU - Adam Willets, 43 field goal BU - David Lang, 21 field goal CU - Adam Willets, 37 field goal RUSHING: BU - Huck, 6-56; Hitchcock, 8-34; Gray, 13-31; Trujillo, 1-28; Koopman, 1-14; Bohrer, 1-6; Team, 1-minus 2. CU - Polk, 1563; Oates, 6-25; Smith, 7-21; Jordan, 3-11; Brown, 2-8; Cramer, 1-5; Kirtz, 2-4; Bryant, 1-3. PASSING: BU - Huck, 22-30-0 for 214 yards. CU - Polk, 15-31-1 for 161 yards. RECEIVING: BU - Watkins, 5-54; Larsen, 3-34; Bohrer, 3-28; Koopman, 3-26; Gray, 3-12; McHale, 2-43; Bork, 2-11; Walley, 1-6. CU - Jordan, 4-66; Stallings, 3-37; Stryffeler, 3-22; Oates, 2-17; Smith, 2-12; Murphy, 1-7. BUTLER 14, DAVIDSON 7 Oct. 31 at Butler Bowl (A - 2,568) Score By Quarters 1 2 3 4 F Davidson 0 0 7 0 7 Butler 7 0 0 7 14 SCORING SUMMARY BU - Ryan Hitchcock, 8 run (David Lang kick) DC - Kenny Mantuo, 9 run (Ben Behrendt kick) BU - Andrew Huck, 1 run (David Lang kick) RUSHING: BU - Gray, 14-117; Huck, 16-81; Hitchcock, 8-58; Bohrer, 2-9; Trujillo, 1-4; Team, 1-minus 1. DC - Mantuo, 19-147; Blanchard, 9-41; Williams, 7-15. PASSING: BU - Huck, 14-26-0 for 136 yards. DC - Blanchard, 1223-1 for 91 yards. RECEIVING: BU - Watkins, 4-33; McHale, 3-55; Bohrer, 2-23; Koopman, 2-20; Hitchcock, 2-minus 7; Larsen, 1-12. DC - Hanabury, 5-40; Aldrich, 4-37; Mantuo, 2-13; Williams, 1-1. BUTLER 31, DAYTON 28 Nov. 7 at Welcome Stadium (A - 4,012) Score By Quarters 1 2 3 4 F Butler 0 13 6 12 31 Dayton 7 0 7 14 28 SCORING SUMMARY UD - Dan Jacob, 4 run (Nick Glavin kick) BU - Andrew Huck, 7 run (David Lang kick) BU - Dan Bohrer, 7 pass from Andrew Huck (Kick failed) BU - Andrew Huck, 7 run (Run failed) UD - Justin Watkins, 50 pass from Steve V Valentino (Glavin kick) BU - Andrew Huck, 36 run (Kick failed) UD - Steve Valentino, 4 run (Pass failed) BU - Carter Walley, 16 pass from Huck (Kick failed) UD - Nick Collins, 8 pass from Valentino (Luke Bellman pass) RUSHING: BU - Gray, 19-109; Huck, 12-54; Koopman, 2-30; Hitchcock, 3-23; Bohrer, 1-14; Trujillo, 1-5. UD - Valentino, 12-44; Mack, 9-39; Jacob, 7-32; Watkins, 1-minus 5; Team, 1-0. PASSING: BU - Huck, 12-29-0 for 146 yards; UD - Valentino, 29-44-2 for 413 yards; Team, 0-1-0. RECEIVING: BU - McHale, 3-21; Bohrer, 2-35; Walley, 2-22; Watkins, 2-9; Hitchcock, 1-35; Koopman, 1-14; Trujillo, 1-10. UD - Watkins, 7-163; Jonard, 7-98; Collins, 7-82; Bellman, 4-51; Mack, 2-12; Papp, 1-9; Millio, 1-minus 2. JACKSONVILLE 36, BUTLER 7 Nov. 14 at Milne Field (A - 1,933) Score By Quarters 1 2 3 4 F Butler 0 7 0 0 7 Jacksonville 0 14 22 0 36 SCORING SUMMARY BU - Dan Bohrer, 7 pass from Andrew Huck (David Lang kick) JU - Christopher Kuck, 10 pass from Josh McGregor (D. Curry kick) JU - Josh Philpart, 42 pass from McGregor (Donovan Curry kick) JU - Christopher Kuck, 2 pass from McGregor (Kuck run) JU - Rudell Small, 54 run (Curry kick) JU - Brian Valdez, 42 pass interception return (Curry kick) RUSHING: BU - Gray, 19-87; Hitchcock, 8-37; Huck, 8-15; Bohrer, 2-11; Trujillo, 1-5. JU - Small, 18-103; McGregor, 5-26; Laster, 2-4; Harris, 2-4 PASSING: BU - Huck, 21-35-2 for 220 yards; Team, 0-1-0. JU - McGregor, 10-19-0 for 157 yards; Stepelton, 1-1 for 4 yards; Curry, 0-1. RECEIVING: BU - Watkins, 8-96; Bohrer, 5-42; Larsen, 4-42; McHale, 2-30; Koopman, 1-5; Trujillo, 1-5. JU - Thompson, 2-65; Philpart, 2-46; Sumter, 2-14; Kuck, 2-12; Laster, 1-12; Davis, 1-8; Louissaint, 1-4. BUTLER 20, DRAKE 17 Nov. 21 at Butler Bowl (A - 2,122) Score By Quarters 1 2 3 4 F Drake 3 0 7 7 17 Butler 0 3 7 10 20 SCORING SUMMARY DU - Brandon Wubs, 23 field goal BU - David Lang, 29 field goal DU - Pat Cashmore, 7 run (Wubs kick) BU - Jordan Koopman, 18 run (David Lang kick) BU - Zach Watkins, 32 pass from Andrew Huck (Lang kick) DU - Joey Orlando, 53 pass from Mike Piatkowski (Wubs kick) BU - David Lang, 27 field goal (0:01) RUSHING: BU - Gray, 15-69; Hitchcock, 7-60; Koopman, 1-18; Trujillo, 1-2; Huck, 6-minus 23; Team, 1-minus 11. DU - Cashmore, 10-39; Piatkowski, 13-26; Morse, 5-7; Broman, 3-5; Platek, 1-3; Kostek, 7-3. PASSING: BU - Huck, 24-35-1 for 234 yards; Team, 0-1. DU - Piatkowski, 17-29-3 for 166 yards. RECEIVING: BU - Watkins, 11-109; McHale, 5-83; Bohrer, 4-38; Trujillo, 2-7; Hitchcock, 1-minus 1; Gray, 1-minus 2. DU - Platek, 6-37; Orlando, 4-70; Pucher, 2-24; Blackmon, 2-23; Kostek, 2-8; Broman, 1-4. BUTLER 28, CENTRAL CONNECTICUT STATE 23 Dec. 5 at Butler Bowl (A - 1,577) Score By Quarters 1 2 3 4 F Central Conn. State 7 0 3 13 23 Butler 0 7 7 14 28 SCORING SUMMARY CC - E. Richardson, 1 run (Joe Izzo kick) BU - Scott Gray, 2 run (David Lang kick) CC - Joe Izzo, 32 field goal BU - Andrew Huck, 10 run (Lang kick) BU - Andrew Huck, 19 run (Lang kick) CC - James Mallory, 5 run (Izzo kick) BU - Andrew Huck, 7 run (Lang kick) CC - E. Richardson, 2 run (Run Failed) RUSHING: BU - Gray, 15-88; Huck, 9-41; Hitchcock, 5-14; Trujillo, 2-9; Bohrer, 1-4; Koopman, 1-minus 5; Team, 1-minus 1. CC - Mallory, 21-109; Fowler, 5-39; Norris, 10-36; Paul, 2-24; Richardson, 6-15; Spadaro, 2-8. PASSING: BU - Huck, 15-21-0 for 182 yards. CC - Norris, 17-26-1 for 202 yards. RECEIVING: BU - Bohrer, 7-68; W Watkins, 4-41; McHale, 2-42; Gray, 1-23; Koopman, 1-7. CC - Paul, 3-57; Grochowski, 3-37; Colagiovanni, 3-37; Fisher, 3-35; Easley, 2-15; Mallory, 2-15; Fowler, 1-6. 23 2010 Opponents Albion Youngstown State Taylor San Diego September 4 Sprankle-Sprandel Stadium Albion, Mich. September 11 Stambaugh Stadium Youngstown, Ohio September 18 Butler Bowl Indianapolis, Ind. September 25 Torero Stadium San Diego, Calif. Location: Albion, Mich. Enrollment: 1,635 Nickname: Britons Colors: Purple and Gold Conference: Michigan Intercollegiate Stadium: Sprankle-Sprandel Stadium (4,244) Athletic Director: Greg Polnasek SID: Bobby Lee Office Phone: (517) 629-0434 Fax: (517) 629-0566 E-Mail: blee@albion.edu Web: Location: Youngstown, Ohio Enrollment: 14,682 Nickname: Penguins Colors: Red and White Conference: Missouri Valley Football Stadium: Stambaugh Stadium (20,630) Athletic Director: Ron Strollo SID: Trevor Parks Office Phone: (330) 941-3192 Fax: (330) 941-3191 E-Mail: tparks@ysu.edu Web: Location: Upland, Ind. Enrollment: 1,975 Nickname: Trojans Colors: Purple and Gold Conference: Mid-States Football Assoc. Stadium: Jim Wheeler Memorial Stadium (4,000) Athletic Director: Dr. Angie Fincannon SID: Eric Smith Office Phone: (765) 998-4569 Fax: (765) 998-4590 E-Mail: ersmith@taylor.edu Web: Location: San Diego, Calif. Enrollment: 7,800 Nickname: Toreros Colors: Torero Blue, Navy and White Conference: Pioneer Football League Stadium: Torero Stadium (7,000) Athletic Director: Ky Snyder SID: Ted Gosen Office Phone: (619) 260-4745 Fax: (619) 260-2990 E-Mail: tgosen@sandiego.edu Web: Head Coach: Craig Rundle Alma Mater: Albion ’74 Career Record: 132-100-1 (24 yrs.) Office Phone: (517) 629-0459 Assistant Coaches: Dustin Beurer, Anthony Cole, Greg Polnasek, D. J. Rehberg, Ron Parker 2009 Record: 4-6 2009 Results (4-6) Butler Thiel At Millikin Central At Olivet Hope Adrian At Trine Kalamazoo At Alma AC OPP 20 6 16 0 13 30 12 13 10 5 21 28 6 28 7 30 23 20 22 34 2010 Schedule Sept. 4 BUTLER Sept. 11 WHEATON Sept. 18 At Greenville Sept. 25 At Wis.-Stevens Point Oct. 2 KALAMAZOO Oct. 9 At Alma Oct. 16 HOPE Oct. 23 At Olivet Oct. 30 ADRIAN Nov. 13 At Trine 24 Head Coach: Eric Wolford Alma Mater: Kansas State ’94 Career Record: First Season Office Phone: (330) 941-3478 Assistant Coaches: Tom Sims, Shane Montgomery, Rick Kravitz, Louie Matsakis, Phil Longo, Frank Buffano, Carmen Bricillo, Andre Coleman, Rollen Smith, Ron Stoops Jr. 2009 Record: 6-5 2009 Results (6-5) At Pittsburgh Austin Peay At Northeastern At Indiana State Missouri State Western Illinois At Southern Illinois South Dakota State At Northern Iowa Illinois State At North Dakota State YSU OPP 3 38 38 21 38 21 28 0 7 17 31 21 8 27 3 17 7 28 30 18 39 35 2010 Schedule Sept. 4 At Penn State Sept. 11 BUTLER Sept. 18 CENT. CONN. STATE Sept. 25 SOUTHERN ILLINOIS Oct. 2 At Missouri State Oct. 9 NORTH DAKOTA STATE Oct. 16 At Western Illinois Oct. 23 At South Dakota State Oct. 30 NORTHERN IOWA Nov. 6 At Illinois State Nov. 13 INDIANA STATE Head Coach: Ron Korfmacher Alma Mater: Taylor ’82 Career Record: 9-11 (2 yrs.) Office Phone: (765) 998-5311 Assistant Coaches: Greg Wolfe, Pete Demorest, Dale Perine, Tony Kijanko, Mike Miley, Lance Brookshire 2009 Record: 7-3 2009 Results (7-3) Anderson At William Penn At St. Francis (Ill.) St. Xavier At Malone Walsh Trinity International At Marian At Olivet Nazarene Saint Francis (Ind.) TU OPP 31 16 16 42 38 23 35 51 23 17 12 31 48 7 36 35 45 35 23 16 2010 Schedule Sept. 2 At Anderson Sept. 11 WILLIAM PENN Sept. 18 At Butler Sept. 25 NOTRE DAME COLLEGE Oct. 2 At St. Xavier Oct. 9 MALONE Oct. 16 At Walsh Oct. 23 At Trinity International Oct. 30 MARIAN Nov. 6 OLIVET NAZARENE Nov. 13 SAINT FRANCIS (IND.) Head Coach: Ron Caragher Alma Mater: UCLA ‘90 Career Record: 22-9 (3 yrs.) Office Phone: (619) 260-4740 Assistant Coaches: Sam Anno, Keith Carter, Tanner Engstrand, Gabe Franklin, Jerome Pathon, Mike Rish, Andrew Rolin, Joe Staab, Jon Sumrall, Dorian Keller 2009 Record: 4-7 2009 Results (4-7) At Azusa Pacific At Northern Colorado Marist At Butler At Valparaiso Drake Jacksonville At Dayton Davidson At Morehead State Southern Utah USD OPP 24 12 12 31 17 10 24 25 48 7 14 21 16 34 14 21 27 34 13 7 32 37 2010 Schedule Sept. 4 AZUSA PACIFIC Sept. 11 At Southern Utah Sept. 18 UC DAVIS Sept. 25 BUTLER Oct. 2 At Jacksonville Oct. 9 DAYTON Oct. 16 At Marist Oct. 23 VALPARAISO Oct. 30 At Drake Nov. 6 MOREHEAD STATE Nov. 13 At Davidson 2010 Opponents Campbell Davidson Dayton Morehead State October 2 Butler Bowl Indianapolis, Ind. October 9 Richardson Stadium Davidson, N.C. October 16 Butler Bowl Indianapolis, Ind. October 23 Butler Bowl Indianapolis, Ind. Location: Buies Creek, N.C. Enrollment: 6,834 Nickname: Fighting Camels Colors: Orange and Black Conference: Pioneer Football League Stadium: CU Football Stadium (5,000) Athletic Director: Stan Williamson SID: Joe Prisco Office Phone: (910) 893-1369 Fax: (910) 893-1330 E-Mail: priscoj@campbell.edu Web: Location: Davidson, N.C. Enrollment: 1,800 Nickname: Wildcats Colors: Red and Black Conference: Pioneer Football League Stadium: Richardson Stadium (4,500) Athletic Director: Jim Murphy SID: Marc Gignac Office Phone: (704) 894-2123 Fax: (704) 894-2636 E-Mail: magignac@davidson.edu Web: Location: Dayton, Ohio Enrollment: 7,700 Nickname: Flyers Colors: Red and Blue Conference: Pioneer Football League Stadium: Welcome Stadium (11,000) Athletic Director: Ted Wabler SID: Doug Hauschild Office Phone: (937) 229-4390 Fax: (937) 229-4461 E-Mail: sid@udayton.edu Web: Location: Morehead, Ky. Enrollment: 9,046 Nickname: Eagles Colors: Blue and Gold Conference: Pioneer Football League Stadium: Jayne Stadium (10,000) Athletic Director: Brian Hutchinson SID: Drew Dickerson Office Phone: (606) 783-2557 Fax: (606) 783-2550 E-Mail: a.dickerson@moreheadstate. edu Web: Head Coach: Dale Steele Alma Mater: South Carolina ‘76 Career Record: 4-18 (2 yrs.) Office Phone: (910) 893-1874 Assistant Coaches: Nick Cavallo, Andre Fontenette, Art Link, Landon Mariani, Oscar Olejniczak, Greg Williams 2009 Record: 3-8 Head Coach: Tripp Merritt Alma Mater: UNC Charlotte ’90 Career Record: 23-28 (5 yrs.) Office Phone: (704) 894-2378 Assistant Coaches: Brett Hayford, Ryan Heasley, Meade Clendaniel, Amish Patel, Jay Poag, Jimmy Means, Robert Wilk, Josh Lustig, Frank Swart, Steve Leahy 2009 Record: 3-7 Head Coach: Rick Chamberlin Alma Mater: Dayton ’80 Career Record: 18-5 (2 yrs.) Office Phone: (937) 229-4423 Assistant Coaches: Dave Whilding, Chris Ochs, Craig Turner, Tony Davis, Patrick Henry, James Stanley, Landon Fox, Kris Ketron, Kevin Hoyng, Trevor Zeiders 2009 Record: 9-2 2009 Results (3-8) CU OPP Methodist 48 28 At Birmingham-So. (OT) 28 35 At Davidson 7 24 At Marist 13 34 Dayton 17 35 At Old Dominion 17 28 Butler 16 23 Morehead State 31 22 At Drake 6 49 At Valparaiso 17 3 Jacksonville 14 34 2009 Results (3-7) At Elon Lenoir-Rhyne Campbell At Jacksonville Morehead State At Dayton Drake At Butler At San Diego Marist 2009 Results (9-2) Urbana At Robert Morris Duquesne At Morehead State At Campbell Davidson At Valparaiso San Diego Butler At Drake Marist 2010 Schedule Sept. 4 At Virginia-Wise Sept. 11 OLD DOMINION Sept. 18 DAVIDSON Sept. 25 GEORGIA STATE Oct. 2 At Butler Oct. 16 DRAKE Oct. 23 At Dayton Oct. 30 MARIST Nov. 6 VALPARAISO Nov. 13 At Jacksonville Nov. 20 At Morehead State 2010 Schedule Sept. 4 GEORGETOWN Sept. 11 At Lenoir-Rhyne Sept. 18 At Campbell Sept. 25 JACKSONVILLE Oct. 9 BUTLER Oct. 16 At Morehead State Oct. 23 At Drake Oct. 30 DAYTON Nov. 6 At Marist Nov. 13 SAN DIEGO Nov. 20 At Presbyterian DC OPP 0 56 0 42 24 7 21 27 16 10 0 17 16 21 7 14 34 27 6 14 UD OPP 10 13 21 14 24 17 30 15 35 17 17 0 36 7 21 14 28 31 23 6 27 16 2010 Schedule Sept. 4 ROBERT MORRIS Sept. 11 At Duquesne Sept. 18 MOREHEAD STATE Sept. 25 CENTRAL STATE Oct. 2 VALPARAISO Oct. 9 At San Diego Oct. 16 At Butler Oct. 23 CAMPBELL Oct. 30 At Davidson Nov. 6 DRAKE Nov. 13 At Marist Head Coach: Matt Ballard Alma Mater: Gardner-Webb ’79 Career Record: 125-109 (22 yrs.) Office Phone: (606) 783-2020 Assistant Coaches: John Gilliam, Gary Dunn, Rob Tenyer, Chris Garner, Paul Humphries 2009 Record: 3-8 2009 Results (3-8) MSU OPP Southern Virginia 61 10 At St. Francis (Pa.) 0 31 At N. Carolina Cent. (2OT) 13 10 Butler (OT) 21 28 Dayton 15 30 At Davidson 10 16 At Jacksonville 0 39 Marist 14 24 At Campbell 22 31 San Diego 7 13 At Valparaiso 29 6 2010 Schedule Sept. 4 At James Madison Sept. 11 ST. FRANCIS (PA.) Sept. 18 At Dayton Sept. 25 At Marist Oct. 2 At Georgia State Oct. 16 DAVIDSON Oct. 23 At Butler Oct. 30 JACKSONVILLE Nov. 6 At San Diego Nov. 13 VALPARAISO Nov. 20 CAMPBELL 25 2010 Opponents Valparaiso Jacksonville Drake Marist October 30 Brown Field Valparaiso, Ind. November 6 Butler Bowl Indianapolis, Ind November 13 Drake Stadium Des Moines, Iowa PFL Member Not on 2010 Butler Football Schedule Location: Valparaiso, Ind. Enrollment: 3,980 Nickname: Crusaders Colors: Brown and Gold Conference: Pioneer Football League Stadium: Brown Field (5,000) Athletic Director: Mark LaBarbera SID: Ryan Wronkowicz Office Phone: (219) 464-5232 Fax: (219) 464-5762 E-Mail: ryan.wronkowicz@valpo.edu Web: Location: Jacksonville, Fla. Enrollment: 3,436 Nickname: Dolphins Colors: Green and Gold Conference: Pioneer Football League Stadium: D. B. Milne Field (5,000) Athletic Director: Alan Verlander SID: Joel Lamp Office Phone: (904) 256-7409 Fax: (904) 256-7424 E-Mail: jlamp@ju.edu Web: Location: Des Moines, Iowa Enrollment: 5,668 Nickname: Bulldogs Colors: Blue and White Conference: Pioneer Football League Stadium: Drake Stadium (14,557) Athletic Director: Sandy Hatfield Clubb SID: Mike Mahon Office Phone: (515) 271-3014 Fax: (515) 271-3015 E-Mail: mike.mahon@drake.edu Web: Location: Poughkeepsie, N.Y. Enrollment: 4,256 Nickname: Red Foxes Colors: Red and White Conference: Pioneer Football League Stadium: Tenney Stadium (5,000) Athletic Director: Tim Murray SID: Mike Ferraro Office Phone: (845) 575-3321 Fax: (845) 471-0466 E-Mail: michael.j.ferraro@marist.edu Web: Head Coach: Dale Carlson Alma Mater: Concordia-Chicago ‘78 Career Record: 110-103-3 (21 yrs.) Office Phone: (219) 464-5513 Assistant Coaches: Bob Muckian, Tony Pierce, Robert Lee, Rob Hansen, Cliff Glover, Marcus Knight, John Dahman 2009 Record: 1-10 Head Coach: Kerwin Bell Alma Mater: Florida ‘87 Career Record: 19-16 (3 yrs.) Office Phone: (904) 256-7470 Assistant Coaches: Jerry Odom, Andy McCleod, Kerry Webb, Ernie Logan, Ernie Mills, Danny Verpaele, Jerry Crafts, Jim Stomps 2009 Record: 7-4 Head Coach: Chris Creighton Alma Mater: Kenyon ‘91 Career Record: 14-8 (2 yrs.) Office Phone: (515) 271-2104 Assistant Coaches: Tim Allen, Rick Fox, Ben Needham, Bill Charles, Jeff Martin, Aaron Selby, Kyle York, Mark Watson 2009 Record: 8-3 Head Coach: Jim Parady Alma Mater: Maine ‘83 Career Record: 96-88-1 (18 yrs.) Office Phone: (845) 575-3699 Assistant Coaches: Scott Rumsey, Tom Kelly, Larry Riley, Bill Roos, Casey Lorenz, Pete Mahoney, Tom Taylor, Nate Fields, Jason Tillery, Clarence Johnson 2009 Record: 7-4 2009 Results (1-10) At St. Joseph’s (Ind.) At Concordia (Wis.) Carthage At Drake San Diego At Butler Dayton At Marist At Jacksonville Campbell Morehead State 2009 Results (7-4) At Webber Int’l At Samford Old Dominion Davidson At Marist Morehead State At San Diego At Drake Valparaiso Butler At Campbell 2009 Results (8-3) Grand View At Marist At South Dakota Valparaiso Missouri S & T At San Diego At Davidson Jacksonville Campbell Dayton At Butler VU OPP 6 31 20 17 24 34 14 34 7 48 14 23 7 38 0 24 20 49 3 17 6 29 2010 Schedule Sept. 2 At Western Illinois Sept. 11 At Franklin Sept. 18 ST. JOSEPH’S (IND.) Sept. 25 DRAKE Oct. 2 At Dayton Oct. 9 MARIST Oct. 16 JACKSONVILLE Oct. 23 At San Diego Oct. 30 BUTLER Nov. 6 At Campbell Nov. 13 At Morehead State 26 JU OPP 40 24 0 27 27 28 27 21 27 31 39 0 34 16 38 45 49 20 36 7 34 14 2009 Schedule Sept. 4 At Old Dominion Sept. 11 At Appalachian State Sept. 18 WEBBER INT’L Sept. 25 At Davidson Oct. 2 SAN DIEGO Oct. 9 DRAKE Oct. 16 At Valparaiso Oct. 23 MARIST Oct. 30 At Morehead State Nov. 6 At Butler Nov. 13 CAMPBELL DU OPP 22 0 34 6 21 51 34 14 19 0 21 14 21 16 45 38 49 6 6 23 17 20 2010 Schedule Sept. 4 LEHIGH Sept. 11 At Missouri S & T Sept. 18 At Montana State Sept. 25 At Valparaiso Oct. 2 MARIST Oct. 9 At Jacksonville Oct. 16 At Campbell Oct. 23 DAVIDSON Oct. 30 SAN DIEGO Nov. 6 At Dayton Nov. 13 BUTLER 2009 Results (7-4) At Sacred Heart Drake At San Diego At Bucknell Campbell Jacksonville At Morehead State Valparaiso Georgetown At Davidson At Dayton MC OPP 31 12 6 34 10 17 16 17 34 13 31 27 24 14 24 0 23 21 14 6 16 27 2010 Schedule Sept. 3 SACRED HEART Sept. 11 BUCKNELL Sept. 25 MOREHEAD STATE Oct. 2 At Drake Oct. 9 At Valparaiso Oct. 16 SAN DIEGO Oct. 23 At Jacksonville Oct. 30 At Campbell Nov. 6 DAVIDSON Nov. 13 DAYTON Nov. 20 At Georgetown Opponents/PFL The Pioneer Football League All-Time Series Records Opponent Akron Ala.-Birmingham Albion Anderson Ashland Austin Peay Ball State Bethany Bradley California-Davis Campbell Centenary Central Conn. State Central State (Ohio) Chicago Cincinnati Clinch Valley Danville Normal Davidson Dayton Denison DePauw Drake Duquesne Earlham Eastern Illinois Evansville Ferris State Findlay Florida International Franklin Georgetown (D.C.) Georgetown (Ky.) George Washington Grand Valley State Hanover Haskell Indians Hillsdale Hofstra Howard Payne Illinois Illinois State Illinois Wesleyan Indiana Indianapolis Indiana State Iowa Jacksonville Kentucky State Lindenwood (Mo.) Lombard Louisville Loyola Manchester Marquette Marshall Miami, Ohio Michigan Michigan State Millikin Minnesota Missouri S & T Moores Hill Morehead State W 0 0 5 2 8 1 22 1 6 0 2 1 1 0 0 3 1 4 1 9 1 44 6 0 19 3 35 1 1 0 36 0 10 1 1 21 1 6 0 2 1 0 1 2 19 27 0 0 3 1 0 2 1 3 0 0 4 0 1 1 0 2 3 3 L 8 2 5 0 5 2 14 1 2 1 0 1 0 1 0 6 1 0 5 25 0 13 12 2 8 3 12 3 0 2 16 1 1 1 4 1 3 4 2 0 8 1 0 5 7 7 1 5 1 1 2 2 1 0 2 1 4 1 1 4 2 2 0 10 T 0 0 0 0 0 0 2 0 0 0 0 0 0 0 1 1 0 0 0 1 0 3 1 0 1 0 2 0 0 0 2 0 0 0 0 1 0 0 0 0 0 0 0 0 2 2 0 0 0 0 0 0 0 0 0 0 1 0 0 0 0 0 0 0 Opponent New York University North Central Northeast Mo. State Northern Michigan Northern Illinois Northwestern Northwood Institute Notre Dame Ohio Ohio Wesleyan Pittsburg State Purdue Quincy (Ill.) Robert Morris Rose Hulman Saginaw Valley San Diego St. Ambrose St. Francis (Ind.) St. Francis (Pa.) St. Joseph’s St. Louis St. Norbert St. Xavier Southern Utah Taylor Tennessee-Martin Thomas More Tiffin Toledo Towson State Transylvania Tufts Valparaiso Wabash Wash. & Jefferson Washington (Mo.) Wayne State Wesley (Del.) Western Kentucky Western Michigan Western Reserve Wilmington Winona Wis.-Stevens Point Wittenberg Xavier Youngstown State W 0 1 1 2 0 0 2 0 1 1 0 3 4 0 8 2 5 2 0 4 34 0 2 1 0 1 0 0 1 2 0 1 1 44 42 1 9 5 2 1 3 2 2 4 2 3 0 0 L 1 0 0 1 2 2 0 3 8 1 1 7 1 4 9 1 12 0 4 0 19 1 1 0 1 0 1 1 3 1 2 1 0 24 17 0 12 1 0 7 10 5 0 0 2 7 3 0 T 0 0 0 0 0 0 0 0 1 0 0 0 0 0 3 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 0 0 8 0 1 0 0 0 0 1 0 0 0 0 0 0 Butler University....................................................Indianapolis, Ind. Campbell University .......................................... Buies Creek, N.C. Davidson College ...................................................Davidson, N.C. University of Dayton....................................................Dayton, Ohio Drake University..................................................Des Moines, Iowa Jacksonville University ....................................... Jacksonville, Fla. Marist College .................................................Poughkeepsie, N.Y. Morehead State University ......................................Morehead, Ky. University of San Diego........................................San Diego, Calif. Valparaiso University..............................................Valparaiso, Ind. The 2010 season marks the 18th year for the Pioneer Football League – the nation’s only non-scholarship NCAA Football Championship Subdivision conference. The league expanded to 10 members a year ago, with Marist College joining as a fulltime member. Currently the PFL plays an eightgame league schedule, rotating the ninth league opponent on a twoyear basis. The PFL is one of only three conferences that sponsor football as its only sport (the Missouri Valley Football Conference and Great West Football Conferences being the others). However, league for the 2008 season. In January 1991 the NCAA passed legislation to require Division I institutions to sponsor all intercollegiate sports at the Division I level. The five charter members (Evansville the fifth classification within Division I and adopted the moniker of Pioneer based on the intent to become the first league in that new division. All 10 current members are committed to the non-scholarship football model. The PFL began play in 1993 with Dayton winning the league’s first crown. The league spent its first season in 1993 under the administrative guidance of the Midwestern Collegiate Conference, and the office moved to St. Louis in 1994 when current commissioner Patty Viverito was named the PFL commissioner, a leadership position she continues to fill. 2009 Pioneer Football League Standings Butler Dayton Drake Jacksonville Marist San Diego Davidson Campbell Morehead State Valparaiso Conference W L T 7 1 0 7 1 0 6 2 0 6 2 0 5 3 0 3 5 0 3 5 0 2 6 0 1 7 0 0 8 0 PCT. .875 .875 .750 .750 .625 .375 .375 .250 .125 .000 W L 11 1 9 2 8 3 7 4 7 4 4 7 3 7 3 8 3 8 1 10 T 0 0 0 0 0 0 0 0 0 0 All Games PCT. Home .917 7-0 .818 4-2 .727 5-1 .636 4-1 .636 4-1 .364 1-4 .300 2-3 .273 2-3 .273 1-4 .091 0-5 Away 4-1 5-0 3-2 3-3 3-3 3-3 1-4 1-5 2-4 1-5 27 Butler Records INDIVIDUAL HONORS 2006 2003 1998 1995 1994 1993 1989 1988 1976 1975 1974 1963 1961 1959 1958 1957 - All-America Chris Marzotto (Football Gazette) Brandon Martin (Football Gazette, Sports Network) Ryan Zimpleman (Sports Network) Arnold Mickens (Sports Network) Arnold Mickens (Sports Network, A. P.) Richard Johnson (Sports Network) Steve Roberts (Kodak) Steve Roberts (Kodak) Bill Lynch (A.P. Little All-America HM) Bill Lynch (A.P. Little All-America HM) Bill Lynch (A.P. Little All-America HM) Lee Grimm (A.P. Little All-America) Don Benbow (A.P. Little All-America) Walt Stockslager (A. P. Little All-America) Paul Furnish (A.P. Little All-America) John Harrell (A.P. Little All-America HM) 2009 1999 1998 1984 - Academic All-America Grant Hunter Mike Goletz Nick Batalis, Mike Goletz Steve Kollias National Scholar Athlete (National Football Foundation Hampshire Honor Society) 2008 - Mike Bennett (National Football Foundation/College Hall of Fame) 1998 - Nick Batalis 2003 1991 1989 1988 1977 1976 1975 1974 - Conference Player Of The Year Brandon Martin (PFL North, Defensive) Paul Romanowski (MIFC) Steve Roberts (HCC) Steve Roberts (HCC) Mike Chrobot (HCC) Bill Lynch (ICC) Bill Lynch (ICC) Bill Lynch (ICC) 1983 1976 1975 1974 1973 1972 1970 1968 1966 1962 1960 1954 - Conference Back Of The Year Paul Romanowski (MIFC Offense) Steve Roberts (HCC Offense) Chuck Orban (HCC Defense) Steve Roberts (HCC Offense) Todd Yeoman (HCC Defense) Eric Chapman (HCC Offense) Bill Lynch (ICC) Bill Lynch (ICC) Bill Lynch (ICC) Keith Himmel (ICC) Noble York (ICC) Dan Nolan (ICC) Larry Gilbert (ICC) Dan Warfel (ICC) Ron Adams (ICC) John Skirchak (ICC) Leroy Thompson (ICC) 1992 1989 1988 1987 1985 1983 1980 1978 1977 1975 1974 1972 1971 1962 1959 1958 1953 - Conference Lineman Of The Year Ruben DeLuna (MIFC Offense) Greg Mariacher (HCC Defense) Mark Allanson (HCC Offense) Todd Jones (HCC Offense) Jeff Palmer (HCC Defense) John Carwile (HCC Offense) Tony Pence (HCC Defense) Mike Chrobot (HCC) Mike Chrobot (HCC) Dave Swihart (ICC) Dave Swihart (ICC) Mike McDevitt (ICC) Tom Redmond (ICC) Lee Grimm (ICC) Jim Ringer (ICC) Paul Furnish (ICC) George Freyn (ICC) 1991 1989 1988 - 2009 2008 2007 2006 2005 - 28 Butler MVP’s Andrew Huck (Offense) Nick Caldicott (Defense) Matt Kobli (Offense) Grant Hunter (Defense) Scott Gray (Offense) Chris Marzotto (Defense) T. J. Brown (Offense) Mike Marzotto (Defense) Rick Tyson (Offense) Chris Marzotto (Defense) Individual MOST GAMES PLAYED Career: 45, Brian Adika, 2006-09 MOST POINTS Game: 36, Steve Roberts vs. St. Ambrose, 11/4/89 Consecutive Games: 54, Scott Gray vs. Albion (30), 9/1/07, and Hanover (24), 9/8/07 Season: 142, Steve Roberts, 1988 Career: 386, Steve Roberts, 1986-89 MOST TOUCHDOWNS Game: 6, Steve Roberts vs. St. Ambrose, 11/4/89 Consecutive Games: 9, Scott Gray vs. Albion (5), 9/1/07, and Hanover (4), 9/8/07 Season: 23, Steve Roberts, 1988 Career: 63, Steve Roberts, 1986-89 MOST EXTRA POINTS Game: 8, Bob Ligda vs. St. Joseph’s, 11/9/74; Tim Witmer vs. St. Ambrose, 11/4/89 Season: 42, Jordan Quiroz, 2008 Career: 114, John Jenkins, 1985-88 MOST FIELD GOALS Game: 5, John Jenkins vs. Dayton, 9/26/87 Season: 11, John Jenkins, 1987; Tim Witmer, 1990 Career: 34, John Jenkins, 1985-88 MOST YARDS RUSHING Game: 295, Arnold Mickens vs. Valparaiso, 10/8/94 Season: 2,255, Arnold Mickens, 1994 Career: 4,623, Steve Roberts, 1986-89 MOST RUSHING ATTEMPTS Game: 56, Arnold Mickens vs. Valparaiso, 10/8/94 Season: 409, Arnold Mickens, 1994 Career: 1,026, Steve Roberts, 1986-89 MOST YARDS PASSING Game: 497, DeWayne Ewing vs. San Diego, 10/21/00 Season: 3,182, DeWayne Ewing, 2000 Career: 8,094, DeWayne Ewing, 1998-2001 MOST PASSES ATTEMPTED Game: 50, Curt Roy vs. St. Joseph’s, 11/7/81; Eli Stoddard vs. San Diego, 10/11/97 Season: 388, DeWayne Ewing, 2000 Career: 1094, DeWayne Ewing, 1998-2001 MOST PASSES COMPLETED Consecutive: 12, DeWayne Ewing vs. Valparaiso, 10/6/01; Matt Kobli at Albion, 9/6/08 Game: 36, DeWayne Ewing vs. St. Francis (Ind.), 9/2/00 Season: 246, DeWayne Ewing, 2000 Career: 656, DeWayne Ewing, 1998-2001 MOST TOUCHDOWN PASSES Game: 5, Ron Kiolbassa vs. Valparaiso, 10/13/90; DeWayne Ewing vs. Quincy, 11/3/01; Andrew Huck vs. Albion, 9/5.09 Season: 23, Matt Kobli, 2008 Career: 60, Bill Lynch, 1972-76 MOST TOTAL OFFENSE Game: 465, DeWayne Ewing vs. San Diego, 10/21/00 Season: 3,079, DeWayne Ewing, 2000 Career: 7,743, DeWayne Ewing, 1998-2001 MOST PASS RECEPTIONS Game: 15, Tom Redmond vs. Ball State, 9/23/72 Season: 78, Zach Watkins, 2009 Career: 192, Dan Bohrer, 2006-09 MOST YARDS RECEIVING Game: 260, Kyle Conner vs. San Diego, 10/21/00 Season: 1,031, Tom Redmond, 1972 Career: 2,241, Dan Bohrer, 2006-09 MOST TOUCHDOWN PASSES CAUGHT Game: 4, Dave Oliver vs. St. Joseph’s, 11/9/74; Dan Bohrer vs. Franklin, 9/13/08 Season: 11, Dan Bohrer, 2008 Career: 22, Dan Bohrer, 2006-09 MOST PUNT RETURNS Game: 9, Eric Voss vs. Valparaiso, 10/19/91; Eric Voss vs. St. Joseph’s, 9/12/92 Season: 42, Noble York, 1972 Career: 103, Eric Voss, 1990-93 MOST PUNT RETURN YARDS Game: 136, Steve Roberts vs. Northwood, 9/9/89 Season: 356, Noble York, 1972 Career: 723, Eric Voss, 1990-93 MOST KICKOFF RETURNS Game: 8, Lou Andreadis vs. San Diego, 11/4/95 Season: 38, Justin Campbell, 2002 Career: 110, Justin Campbell, 2001-04 MOST KICKOFF RETURN YARDS Game: 243, Justin Campbell vs. San Diego, 10/19/02 Season: 985, Justin Campbell, 2002 Career: 2,512, Justin Campbell, 2001-04 MOST PUNTS Game: 14, Ron Stryzinski vs. Wittenberg, 9/25/82 Season: 80, Ron White, 1990 Career: 255, Ron White, 1990-93 MOST YARDS PUNTING Game: 493, Ron Stryzinski vs. Wittenberg, 9/25/82 Season: 3,122, Ron White, 1990 Career: 9,731, Ron White, 1990-93 BEST PUNTING AVERAGE Season: 43.2, Shawn Wood, 1998 Career: 39.5, Shawn Wood, 1995-98 MOST SOLO TACKLES Game: 21, Kevin Johnson vs. Ashland, 11/2/91 Season: 101, Chuck Orban, 1990 Career: 269, Chuck Orban, 1987-90 MOST ASSISTED TACKLES Game: 16, John Doctor vs. Ashland, 10/23/82; Joe Miles vs. Dayton, 10/4/97 Season: 81, Chuck Orban, 1989 Career: 236, Dave Ginn, 1981-84 MOST TOTAL TACKLES Game: 29, Kevin Johnson vs. Ashland, 11/2/91; Kevin Johnson vs. Pittsburg State, 11/26/91 Season: 181, Chuck Orban, 1990 Career: 487, Chuck Orban, 1987-90 Butler Records MOST QUARTERBACK SACKS Game: 5, Elgin Reese vs. St. Joseph’s, 9/12/92 Season: 17, Scott Cook, 1983 Career: 39, Tony Pence, 1978-81 MOST FUMBLE RECOVERIES Game: 3, Chuck Orban vs. Dayton, 9/24/88 Season: 4, Steve Torrence, 1980; Chuck Orban, 1988; Bob Espich, 1988 Career: 8, Steve Torrence, 1980-83 MOST PASSES COMPLETED Game: 36 vs. St. Francis (Ind), 9/2/00 Season: 250, 2000 2004 - MOST PASSES, TWO TEAMS Game: 96, Butler (44) vs. Franklin (52), 10/9/82 2002 - MOST TOTAL OFFENSE Game: 673 vs. St. Ambrose, 11/4/89 Season: 4,644, 2009 2000 - MOST TOTAL PLAYS Game: 94 vs. Franklin, 10/26/85 Season: 804, 2009 LONGEST PLAYS Run: 96, Harry Muta vs. Valparaiso, 10/11/75; Don Kelly vs. Indiana State, 11/17/51 Pass: 88, John Moses to John Harrell vs. Washington (Mo.), 9/17/56; Mike McGeorge to Tom Wallace vs. Ashland, 10/18/80 Field Goal: 57, Bob Ligda vs. St. Joseph’s, 10/16/76 Punt: 72, Mike Morgan vs. Morehead State, 10/20/07 Kickoff Return: 98, D’Andre Dowdy vs. Valparaiso, 10/15/05 Punt Return: 84, Noble York vs. Ball State, 9/23/72 Int. Return: 89, Derek Bradford vs. Valparaiso, 10/18/08 Fumble Return: 86, Buck Ulery vs. Dayton, 10/23/04 MISCELLANEOUS TEAM Best Record: 9-0, 1959, 1961 Most Wins: 11, 2009 Consecutive Wins, One Season: 9, 1959, 1961, 2009 Consecutive Wins, Multiple Seasons: 16, 1960-62 Consecutive Games Without A Loss: 19, 1960-62 Most Shutouts: 6, 1905, 1922, 1937 Most Shutouts (Modern Record): 3, 1947, 1958 Consecutive Shutouts: 6, 1937 Consecutive Shutouts (Modern Record): 2, 1947 Team MOST POINTS Game: 122 vs. Hanover, 1921 Game (Modern Record): 68 vs. Indianapolis, 1948 Season: 356, 2000 By Butler Opponent: 81, Minnesota, 1926 By Butler Opp. (Modern Record): 65, Ball State, 1967 Butler + Opp.: 122, Butler (122) vs. Hanover (0), 1921 Butler + Opp. (Modern Record): 113, Butler (56) vs. Georgetown (57) - OT, 9/23/00 In a Loss: 56, Butler vs. Georgetown (57), 9/23/00 2003 - 2001 - 1999 1998 - MOST PASS INTERCEPTIONS Game: 3, George Yearsich vs. St. Joseph’s, 10/16/71; Dan Fuhs vs. Indianapolis, 11/8/80; Weylin Stewart vs. Grand Valley State, 9/21/91; Brandon Martin vs. Davidson, 9/19/03 Season: 9, Noble York, 1972 Career: 17, Brandon Martin, 2000-03 Note: An item carried in Believe It or Not by Ripley noted Kenneth Booz of Butler University punted 110 yards against Ft. Harrison on Oct. 10, 1928. Butler MVP’s (Continued) MOST FIRST DOWNS Game: 31 vs. Valparaiso, 10/1/83; UAB, 11/6/93 Season: 255, 2009 1997 1996 1995 1994 1993 1992 1991 1990 1989 1988 1987 1986 1985 - The Last Time... 1984 1983 - Interception Returned For A Touchdown Butler: Andy Dauch (19 yds.) vs. Hanover, 9/19/09 Opp.: Brian Valdez (42 yds.), Jacksonville, 11/14/09 Fumble Returned For A Touchdown Butler: Collin McGann vs. Robert Morris, 66 yds, 9/16/06 Opp.: Shaun Lewis, Jacksonville, 38 yards, 11/1/08 Punt Returned For A Touchdown Butler: Tadd Dombart, 75 yds vs. Morehead St., 10/25/08 Opp.: Matt Loula, Missouri-Rolla, 5 yds, 9/22/07 Kickoff Returned For A Touchdown Butler: D’Andre Dowdy, 98 yards vs. Valparaiso, 10/15/05 Opp.: Jason Jones, Drake, 73 yards, 10/16/04 200-Yard Rushing Game Butler: Naim Sanders (244) vs. St. Francis (Pa.), 9/19/98 Opp.: Justin Williams (225), Davidson, 10/27/07 1982 1981 1980 1979 1978 1977 1976 1975 1974 1973 1972 1971 1970 - MOST YARDS RUSHING Game: 406 vs. Valparaiso, 10/15/88 Season: 2,544, 1975 100-Yard Receiving Game Butler: Zach Watkins (109) vs. Drake, 11/21/09 Opp.: Justin Watkins (163), Dayton, 11/7/09 1969 - MOST RUSHING ATTEMPTS Game: 80 vs. Franklin, 10/23/76 Season: 531, 1959 300-Yard Passing Game Butler: Andrew Huck (316) vs. Franklin, 9/12/09 Opp.: Steve Valentino (413), Dayton, 11/7/00 1966 - MOST YARDS PASSING Game: 497 vs. San Diego, 10/21/00 Season: 3,233, 2000 Safety Scored Butler: vs. Jacksonville, 11/1/08 Opp.: Drake, 10/6/07 MOST PASSES ATTEMPTED Game: 54 vs. Ball State, 9/23/72 Season: 404, 2000 Defensive 2-Point Conversion Butler: Jacob Fritz vs. Davidson, 10/27/07 Opponent: Tony Tobias, Lindenwood, 11/14/98 1968 1967 - 1965 1964 1963 1962 1961 1960 1959 1958 1957 1956 1955 1954 - Billy Nardini (Offense) Jim Hart (Defense) Bob Leonard (Offense) Brandon Martin (Defense) Dale Jennings (Offense) Lou Capizzi (Defense) DeWayne Ewing (Offense) Nick Ober (Defense) DeWayne Ewing (Offense) Nick Ober (Defense) David Bailey (Offense) Steve Hall (Defense) Nick Batalis, Naim Sanders (Offense) Kevin Ward (Defense) Eli Stoddard, Jeremy Harkin (Offense) Kevin Ward (Defense) Naim Sanders (Offense) Ron Griswold (Defense) Arnold Mickens (Offense) Adam Borrell (Defense) Arnold Mickens (Offense) Chris Toner, Cameron McDaniel (Defense) Richard Johnson (Offense) Dave Kathman (Defense) Kevin Kimble, Ruben DeLuna (Offense) Don DeCraene (Defense) Paul Romanowski (Offense) Kevin Johnson (Defense) Todd Roehling (Offense) Chuck Orban (Defense) Steve Roberts (Offense) Chuck Orban (Defense) Steve Roberts (Offense) Todd Yeoman (Defense) Steve Roberts (Offense) Joe Annee (Defense) Paul Page (Offense) Mark Ribordy (Defense) Paul Page (Offense) Ron Bunt (Defense) Jim Hoskins (Offense) Dave Ginn (Defense) Curt Roy (Offense) Steve Torrence (Defense) Andy Howard, Dave Newcomer (Offense) Dave Ginn, Chris McGary (Defense) Curt Roy (Offense) Tony Pence (Defense) Andy Howard (Offense) Tony Pence (Defense) Ken LaRose (Offense) Paul Harrington (Defense) Mike Chrobot (Offense) Mike Daughty (Defense) Bruce Scifres (Back) Kevin Greisl (Lineman) Bill Lynch (Back) Bruce Ford (Lineman) Bill Lynch (Back) Dave Swihart (Lineman) Bill Lynch (Back) Andy Wetzel (Lineman) Steve Clayton (Back) Phil Schluge (Lineman) Tom Redmond (Back) Ron Cooper, Fred Powell (Lineman) Leonard Brown (Back) Phil Fitzsimmons, Ephraim Smiley (Lineman) Dan Nolan (Back) Mike Caito (Lineman) Dick Reed (Back) Mike Caito (Lineman) Dick Reed (Back) Steve Orphey (Lineman) Mike Harrison (Back) Vic Wukovits (Lineman) Dan Warfel (Back) Larry Fairchild (Lineman) Dick Dullaghan Dick Dullaghan Lee Grimm Ron Adams Phil Long John Skirchak Bob Stryzinski Paul Furnish John Harrell Robert Eichholtz Leroy Thompson, Jim Baker Les Gerlach 29 All-Time Butler Leaders Single Game DeWayne Ewing Rushing Yards 2,255, Arnold Mickens, 1994 1,558, Arnold Mickens, 1995 1,535, Richard Johnson, 1993 1,490, Steve Roberts, 1987 1,450, Steve Roberts, 1989 1,427, Steve Roberts, 1988 1,284, Naim Sanders, 1996 1,190, Kevin Kimble, 1992 1,175, Kevin Kimble, 1991 1,163, Naim Sanders, 1998 Rushing Yards 4,623, Steve Roberts, 1986-89 3,813, Arnold Mickens, 1994-95 3,666, Naim Sanders, 1995-98 2,879, Leroy Thompson, 1953-56 2,802, Kevin Kimble, 1989-92 2,691, Andy Howard, 1979-82 2,445, Harry Muta, 1972-75 2,013, Bruce Scifres, 1975-78 1,920, Richard Johnson, 1990-93 1,889, Kevin McDevitt, 1973-76 Rushing Attempts 56, Arnold Mickens vs. Valparaiso, 1994 54, Arnold Mickens vs. Dayton, 1994 47, Arnold Mickens vs. San Diego, 1994 46, Arnold Mickens vs. UW-Stevens Pt., ‘94 45, Arnold Mickens vs. Drake, 1994 44, Arnold Mickens vs. Georgetown, 1994 44, Steve Roberts vs. Dayton, 1988 44, Steve Roberts vs. Indianapolis, 1989 44, Naim Sanders vs. Evansville, 1996 43, Arnold Mickens vs. UAB, 1994; Howard Payne, 1995; Valparaiso, 1995 Rushing Attempts 409, Arnold Mickens, 1994 354, Arnold Mickens, 1995 333, Steve Roberts, 1987 325, Steve Roberts, 1989 322, Richard Johnson, 1993 315, Steve Roberts, 1988 307, Kevin Kimble, 1991 288, Naim Sanders, 1996 275, Kevin Kimble, 1992 268, Kevin McDevitt, 1976 Rushing Attempts 1,026, Steve Roberts, 1986-89 843, Naim Sanders, 1995-98 763, Arnold Mickens, 1994-95 754, Andy Howard, 1979-82 701, Kevin Kimble, 1989-92 581, Bruce Scifres, 1975-78 479, Eric Chapman, 1981-84 478, Harry Muta, 1972-75 460, Kevin McDevitt, 1973-76 415, Richard Johnson, 1990-93 Passing Yards 3,182, DeWayne Ewing, 2000 2,518, Matt Kobli, 2008 2,454, Andrew Huck, 2009 2,337, DeWayne Ewing, 2001 2,214, Paul Romanowski, 1991 2,179, Mike Lee, 1986 2,118, Eli Stoddard, 1997 2,023, Rob Cutter, 1985 2,012, Ron Kiolbassa, 1990 1,994, Curt Roy, 1983 Passing Yards 8,094, DeWayne Ewing, 1998-2001 5,909, Bill Lynch, 1972-76 4,830, Steve Clayton, 1970-73 4,561, Curt Roy, 1979-83 4,071, Rob Cutter, 1984-87 3,790, Jason Stahl, 1992-94 3,735, Ron Kiolbassa, 1987-90 3,527, Mike Lee, 1982-86 3,170, Dick Reed, 1966-69 2,995, Eli Stoddard, 1994-97 Receiving Yards 1,031, Tom Redmond, 1972 998, Paul Page, 1986 971, Dan Bohrer, 2008 918, Zach Watkins, 2009 876, Kyle Conner, 2000 875, Paul Page, 1985 818, Dave Oliver, 1975 760, Adam Lafferty, 2001 756, Jim Hoskins, 1984 747, John Barron, 1987 Receiving Yards 2,241, Dan Bohrer, 2006-09 2,176, Eric Voss, 1990-93 2,070, Adam Lafferty, 1999-2002 1,937, Paul Page, 1983-86 1,933, Tom Redmond, 1970-72 1,801, Kyle Conner, 1997-2000 1,759, Jon Hill, 1990-93 1,702, Mike Chrobot, 1975-78 1,634, Dave Swihart, 1972-75 1,566, Tom Wallace, 1979-82 Receptions 78, Zach Watkins, 2009 73, Tom Redmond, 1972 71, Dan Bohrer, 2008 63, Kyle Conner, 2000 55, Paul Page, 1985 55, Mike Chrobot, 1978 54, Dave Oliver, 1975 53, Todd Roehling, 1990 52, Dan Bohrer, 2009 52, Nate Miller, 2006 52, Paul Page, 1986 Receptions 192, Dan Bohrer, 2006-09 142, Eric Voss, 1990-93 137, Mike Chrobot, 1975-78 124, Kyle Conner, 1997-2000 120, Steve Roberts, 1986-89 117, Tom Redmond, 1970-72 117, Paul Page, 1983-86 112, Adam Lafferty, 1999-2002 112, Dave Swihart, 1972-75 111, Jon Hill, 1990-93 Receiving Yards 260, Kyle Conner vs. San Diego, 2000 223, Todd Roehling vs. Dayton, 1990 211, Tom Redmond vs. Ball State, 1972 208, Eric Voss vs. Valparaiso, 1990 183, Zach Watkins vs. Albion, 2009 181, Tom Wallace vs. Georgetown, 1981 176, Paul Page vs. Dayton, 1986 175, Jim Hoskins vs. Franklin, 1984 169, Nate Miller vs. Valparaiso, 2006 165, John Barron vs. Grand Valley State, ‘87 Receptions 15, Tom Redmond vs. Ball State, 1972 13, Zach Watkins vs. Morehead St., 2009 13, Mike Chrobot vs. Dayton, 1978 12, Todd Roehling vs. Dayton, 1990 11, Zach Watkins vs. Drake, 2009 11, Nate Miller vs. Hanover, 2006 11, Kyle Conner vs. San Diego, 2000 11, Kyle Conner vs. Morehead State, 2000 11, Kyle Conner vs. Albion, 1999 11, Steve Roberts vs. Dayton, 1989 Scoring 36, Steve Roberts vs. St. Ambrose, 1989 30, Scott Gray vs. Albion, 2007 24, Dan Bohrer vs. Franklin, 2008 24, Scott Gray vs. Hanover, 2007 24, Dale Jennings vs. UW-Stevens Pt., 2002 24, Dale Jennings vs. Drake, 2002 24, Naim Sanders vs. Valparaiso, 1996 24, Arnold Mickens vs. Valparaiso, 1995 24, Steve Roberts vs. Northwood, 1989 24, Steve Roberts vs. St. Joseph’s, 1989 24, Steve Roberts vs. St. Joseph’s, 1988 24, Steve Roberts vs. Kentucky State, 1988 24, Steve Roberts vs. Northwood, 1988 24, Steve Roberts vs. Franklin, 1987 24, Dave Oliver vs. St. Joseph’s, 1974 24, Kevin McDevitt vs. Franklin, 1976 Steve Roberts 30 Career Rushing Yards 295, Arnold Mickens vs. Valparaiso, 1994 293, Arnold Mickens vs. Drake, 1994 288, Arnold Mickens vs. UW-Stevens Pt., ‘94 271, Steve Roberts vs. St. Ambrose, 1989 263, Arnold Mickens vs. San Diego, 1994 259, Arnold Mickens vs. Valparaiso, 1995 258, Steve Roberts vs. Indianapolis, 1989 251, Naim Sanders vs. Evansville, 1996 246, Steve Roberts vs. N.E. Missouri, 1987 244, Arnold Mickens vs. Evansville, 1994 Passing Yards 497, DeWayne Ewing vs. San Diego, 2000 419, Travis Delph vs. UW-Stevens Pt., 2002 384, Paul Romanowski vs. Hillsdale, 1991 364, Curt Roy vs. St. Joseph’s, 1981 350, DeWayne Ewing vs. St. Francis (Ind.), ‘00 349, DeWayne Ewing vs. Albion, 2001 346, DeWayne Ewing vs. Georgetown, 2000 341, DeWayne Ewing vs. Dayton, 2000 331, Rob Cutter vs. Georgetown, 1985 329, Ron Kiolbassa vs. Kentucky State, ‘89 Arnold Mickens Single Season Scoring 142, 132, 120, 108, 104, 100, 90, 84, 84, 78, 72, Steve Roberts, 1988 Steve Roberts, 1989 Dale Jennings, 2002 Arnold Mickens, 1994 Steve Roberts, 1987 Kevin McDevitt, 1976 Scott Gray, 2007 Ryan Zimpleman, 2000 John Skirchak, 1960 Harry Muta, 1975 Naim Sanders, 1996 Scoring 386, 216, 202, 186, 176, 175, 172, 162, 162, 161, Steve Roberts, 1986-89 John Jenkins, 1985-88 Leroy Thompson, 1953-56 Naim Sanders, 1995-98 Arnold Mickens, 1994-95 Jordan Quiroz, 2005-08 Kevin McDevitt, 1973-76 Dale Jennings, 1998-2002 Ryan Zimpleman, 1997-2000 Tim Witmer, 1989-92 Butler Records BUTLER’S FIRST TEAM ALL-CONFERENCE PLAYERS Indiana Collegiate Conference 1952: Fred Davis, QB; Ralph London, T; John Riddle, HB; Bill Norkus, C; Norm Ellenberger, FB; Don Kelly, HB 1953: Fred Davis, QB; Bob Eichholtz, G; George Freyn, E; Gene Kuzmic, HB; Ralph London, T; Leroy Thompson, FB 1954: Ralph London, T; Leroy Thompson, FB 1956: Bob Eichholtz, G; Ken Nicholson, T; Leroy Thompson, FB 1957: Paul Furnish, G; Phil Mercer, HB 1958: Paul Furnish, G; John Moses, QB; Cliff Oilar, HB; Ken Spraetz, E; Kent Stewart, FB; Bob White, C 1959: Jim Ringer, C; Walt Stockslager, T 1960: John Skirchak, HB; Ed McCauley, G; Don Benbow, T; Phil Long, QB; Gary Green, FB 1961: Phil Long, QB; Don Benbow, T; Larry Shook, HB 1962: Tim Renie, E; Lee Grimm, G; Ron Adams, QB; John Brown, HB 1963: Lee Grimm, G; Rich Florence, E 1965: Dick Dullaghan, HB; Ken Leffler, C; Tom Sayer, T 1966: Larry Fairchild, T; Mark Steinmetz, G; Dan Warfel, LB 1968: Howard Cline, C; Eddie Bopp, DB; Pat Kress, T; Larry Gilbert, HB 1969: Rich Gray, DB; Phil Whisner, DT; Andy Carson, C, Dick Reed, QB 1970: Dan Nolan, HB 1971: Tom Redmond, E, Mike McDevitt, C 1972: Mike McDevitt, C; Noble York, DB, Dave Swihart, TE; Andy Wetzel, G; Tom Redmond, FL; Bob Rykovich, E; Fred Powell, T; Ron Cooper, MG 1973: Dave Swihart, TE; Phil Schluge, T; Steve Clayton, QB; Bill Kuntz, DE; Keith Himmel, DB 1974: Dave Swihart, TE; Bill Lynch, QB; Andy Wetzel, T; Bob Ligda, K; Tim O’Banion, E; Lee Schluge, T 1975: Dave Swihart, TE; Bill Lynch, QB; Harry Muta, RB; Dave Oliver, FL; Bob Ligda, K; Rob Goshert, DE; Dave Cunningham, C; Mark Chappuis, LB 1976: Bill Lynch, QB; Kevin McDevitt, FB Heartland Collegiate Conference 1977: Mike Chrobot, TE; Ken LaRose, T; Chuck Schwanekamp, G; Joe Chaulk, C; Bruce Scifres, HB; Bill Ginn, FB; Ed Thompson, K; Rob Goshert, DE; Mike Minczeski, LB; Bruce Ford, T 1978: Mike Chrobot, TE; Scot Shaw, DB; Mike Daugherty, DE 1979: Ken LaRose, T; Tony Pence, MG; Paul Harrington, LB; Mike Shibinski, DB 1980: Andy Howard, RB; Tim Flanigan, OG; Tony Pence, MG 1981: Landy Breeden, DE; Tony Pence, MG 1982: Dave Newcomer, OT; Landy Breeden, DE; Dave Ginn, LB; Mike Peconge, MG; Tony Sales, DB 1983: Jim Bell, OG; Brian Bertke, MG; John Carwile, C; Eric Chapman, RB; Scott Cook DT; Dave Ginn, LB; John Doctor, LB; Dino Merlina, DB; Scott Olinger, OT; Noble Parks, FB; Curt Roy, QB; John Warne, TE; Steve Torrence, DE 1984: Jim Hoskins, WR; Scott Olinger, OT; Jay Barnhorst, FB; Dave Ginn, LB; Dino Merlina, DB 1985: Paul Page, WR; Jay Barnhorst, FB; Mark Ribordy, DE; Ron Bunt, LB; Jeff Palmer, DT 1986: Paul Page, WR; Mark Ribordy, DE; Todd Jones, OT; George Dury, C; Mike Hegwood, DB 1987: Todd Jones, OT; Rusty Melzoni, OG; Mark Allanson, C; Steve Roberts, RB; John Jenkins, K; Jack Fillenwarth, DE; Tom Klusman, MG; Joe Annee, LB; Todd Yeoman, DB 1988: John Barron, WR; Tom Maheras, OT; Dan Shirey, OG; Mark Allanson, C; Steve Roberts, RB; Randy Renners, DT; Ron Baird, DE; Chuck Orban, LB; Todd Yeoman, DB 1989: Dan Shirey, OG; Mark Allanson, C; Ron Roembke, TE; Steve Roberts, RB; Greg Mariacher, DT; Jerry Pianto, MG; Chuck Orban, LB; Dax Gonzalez, DB; Kevin Shomber, P Midwest Intercollegiate Football Conference 1990: Kevin Enright, OT; Jerry Pianto, NG; Chuck Orban, LB 1991: Todd Larson, OT; Paul Romanowski, QB; Kevin Kimble, RB; Dave Kathman, DE; Kevin Johnson, LB; Dax Gonzalez, DB 1992: Ruben DeLuna, OT; Kevin Kimble, RB; Dave Kathman, LB; Don DeCraene, DB Pioneer Football League 1993: Eric Voss, WR; Jeff Burks, OT; Richard Johnson, RB; Terry Bolen, DT; Dave Kathman, LB 1994: Damon Black, OT; Keith Rossell, OT; Arnold Mickens, RB; Brian Sanders, DT; Marty Erschen, LB; Chris Toner, LB; Cameron McDaniel, DB 1995: Arnold Mickens, RB; Ron Griswold, NG 1996: Naim Sanders, RB; Ron Griswold, NG; Nick Winings, LB 1997: Nick Batalis, C; Naim Sanders, RB; Ron Griswold, NG; Kevin Ward, LB 1998: Nick Batalis, C; Naim Sanders, RB; Adam Timm, WR; Kevin Ward, LB; Shawn Wood, P 1999: Kyle Conner, WR 2000: Kyle Conner, WR; Brandon Willett, TE; Grant Veith, OL; DeWayne Ewing, QB 2001: Kyle Derickson, WR; Brandon Willett, TE; Roman Speron, RB; Nick Ober, LB 2002: Carl Erickson, OL; Dale Jennings, RB; Tim Lytle, DL; Parker Smith; LB, Russ Mann, DB; Justin Campbell, RS 2003: Brandon Martin, DB; John Leininger, DL; Justin Campbell, RS 2004: Dave McMahon, DL; Andy Nelson, DB 2005: Chris Marzotto, DL 2006: Chris Marzotto, DL 2008: Grant Hunter, DL 2009: Spencer Summerville, DB; Donnie Gilmore, OL; Scott Gray, RB; Zach Watkins, WR Conference Champions Indiana Collegiate Conference 1951 Valparaiso (4-0) 1952 BUTLER, Valparaiso (3-1-1) 1953 BUTLER (5-0) 1954 Valparaiso (5-1) 1955 St. Joseph’s, Evansville (5-1) 1956 St. Joseph’s (6-0) 1957 St. Joseph’s (6-0) 1958 BUTLER (5-1) 1959 BUTLER (6-0) 1960 BUTLER (5-1) 1961 BUTLER (6-0) 1962 BUTLER (4-1-1) 1963 BUTLER (6-0) 1964 BUTLER, Ball State, Evansville, Valparaiso, Indiana State (4-2) 1965 Ball State (6-0) 1966 Ball State (5-0-1) 1967 Ball State (5-1) 1968 Valparaiso (4-0) 1969 Valparaiso, Evansville (3-1) 1970 Evansville (4-0) 1971 St. Joseph’s (4-0) 1972 BUTLER, Evansville (4-1) 1973 BUTLER (4-1) 1974 BUTLER (6-0) 1975 BUTLER (6-0) 1976 Evansville, St. Joseph’s (4-1) Heartland Collegiate Conference 1977 BUTLER, St. Joseph’s (3-1) 1978 Indianapolis (4-2) 1979 St. Joseph’s (4-1) 1980 Ashland, Franklin (5-2) 1981 Franklin, Indianapolis (6-1) 1982 Ashland (6-1) 1983 BUTLER (5-0-1) 1984 Ashland (6-1) 1985 BUTLER, Ashland (5-1) 1986 Ashland (5-1) 1987 BUTLER (4-0-1) 1988 BUTLER (3-0-1) 1989 BUTLER (4-0) Midwest Intercollegiate Football Conference 1990 Grand Valley State (9-1) 1991 BUTLER (9-1) 1992 BUTLER, Ferris State, Grand Valley State (8-2) Pioneer Football League 1993 Dayton (5-0) 1994 BUTLER (4-1) Dayton (4-1) 1995 Drake (5-0) 1996 Dayton (5-0) 1997 Dayton (5-0) 1998 Drake (4-0) 1999 Dayton (4-0) 2000 Dayton, Drake, Valparaiso (3-1) 2001 Dayton (4-0) 2002 Dayton (4-0) 2003 Valparaiso (3-1) 2004 Drake (4-0) 2005 San Diego (4-0) 2006 San Diego (7-0) 2007 Dayton (6-1), San Diego (6-1) 2008 Jacksonville (7-1) 2009 Butler, Dayton (7-1) 31 Season-By-Season Records YEAR 32 RECORD COACH 3-0-0 Clint Howe (No football played) 2-0-0 Clint Howe 3-0-1 Clint Howe 4-3-0 (Coach Unknown) 5-1-0 (Coach Unknown) 1-0-0 (Coach Unknown) 6-1-0 Joseph Flint 2-2-0 Joseph Flint 4-3-0 (Coach Unknown) 3-0-0 James Zink 2-1-1 T. Williamson 1-3-0 Walter Kelly 1-3-0 Walter Kelly 1-0-0 Walter Kelly 1-3-0 Walter Kelly 0-3-0 Walter Kelly 6-1-0 R. Wingard 7-2-1 R. Wingard 1-0-0 Don Robinson 1-3-2 John McKay 5-0-1 John McKay 5-3-0 Walter Gipe 4-3-1 John McKay 4-3-1 Dave Allerdice 5-3-0 G. Cullen Thomas 2-4-1 G. Cullen Thomas 4-2-0 G. Cullen Thomas 1-6-0 G. Cullen Thomas 3-5-0 G. Cullen Thomas 3-3-0 G. Cullen Thomas 2-1-1 G. Cullen Thomas 0-5-1 Joe Mullane 7-1-0 Pat Page 6-2-0 Pat Page 8-2-0 Pat Page 7-2-0 Pat Page 4-5-0 Pat Page 5-2-2 Pat Page 3-6-0 Paul D. (Tony) Hinkle 4-3-1 George Potsy Clark 6-2-0 George Potsy Clark 4-4-0 George Potsy Clark 2-7-0 Harry M. Bell 3-5-0 Harry M. Bell 2-4-1 Fred Mackey 2-6-0 Fred Mackey 6-1-1 Fred Mackey 7-1-0 Paul D. (Tony) Hinkle 6-0-2 Paul D. (Tony) Hinkle 5-2-1 Paul D. (Tony) Hinkle 4-4-0 Paul D. (Tony) Hinkle 7-0-1 Paul D. (Tony) Hinkle 4-4-1 Paul D. (Tony) Hinkle 5-4-0 Paul D. (Tony) Hinkle 2-7-0 Frank (Pop) Hedden (No football–World War II) (No football–World War II) 3-3-0 Frank (Pop) Hedden 7-1-0 Paul D. (Tony) Hinkle 5-3-1 Paul D. (Tony) Hinkle 3-5-0 Paul D. (Tony) Hinkle 2-6-0 Paul D. (Tony) Hinkle 4-4-1 Paul D. (Tony) Hinkle 4-4-1 Paul D. (Tony) Hinkle 5-3-1 Paul D. (Tony) Hinkle 6-2-0 Paul D. (Tony) Hinkle 4-4-1 Paul D. (Tony) Hinkle YEAR 1955 1956 1957 1958 1959 1960 1961 1962 1963 1964 1965 1966 1967 1968 1969 1970 1971 1972 1973 1974 1975 1976 1977 1978 1979 1980 1981 1982 1983 RECORD 3-5-0 6-2-0 7-2-0 8-1-0 9-0-0 8-1-0 9-0-0 5-2-2 8-1-0 4-4-1 6-3-0 4-5-0 2-7-0 2-7-0 3-6-0 3-6-1 3-7-0 5-5-0 5-5-0 8-2-0 9-1-0 6-4-0 5-5-0 5-5-0 5-5-0 5-5-0 3-7-0 7-3-0 9-1-1 “Tony” Hinkle COACH YEAR 1984 1985 1986 1987 1988 1989 1990 1991 1992 1993 1994 1995 1996 1997 1998 1999 2000 2001 2002 2003 2004 2005 2006 2007 2008 2009 Summary: Bill Lynch RECORD 6-4-0 8-2-0 5-5-0 8-1-1 8-2-1 7-2-1 5-5-1 9-2-0 8-2-0 4-6-0 7-3-0 2-8-0 3-7-0 6-4-0 4-6-0 5-5-0 2-8-0 5-5-0 4-6-0 2-9-0 1-10-0 0-11-0 3-8-0 4-7-0 6-5-0 11-1-0 Years 120 COACH Bill Sylvester Bill Lynch Bill Lynch Bill Lynch Bill Lynch Bill Lynch Bob Bartolomeo Bob Bartolomeo Ken LaRose Ken LaRose Ken LaRose Ken LaRose Ken LaRose Ken LaRose Ken LaRose Ken LaRose Ken LaRose Ken LaRose Kit Cartwright Kit Cartwright Kit Cartwright Kit Cartwright Jeff Voris Jeff Voris Jeff Voris Jeff Voris Won Lost Tied 539 417 35 Bob Bartolomeo Ken LaRose BUTLER FOOTBALL COACHING RECORDS Coach (Years) Clint Howe (1887-1890) Joseph Flint (1894-1895) James Zink (1897) T. Williamson (1898) Walter Kelly (1899-1903) R. Wingard (1904-1905) Don Robinson (1906) John McKay (1907-1908,1910) Walter Gipe (1909) Dave Allerdice (1911) G. Cullen Thomas (1912-1918) John Mullane (1919) Harland (Pat) Page (1920-1925) George Potsy Clark (1927-1929) Harry Bell (1930-1931) Fred Mackey (1932-1934) Frank (Pop) Hedden (1942-1945) Paul D. (Tony) Hinkle (1926, 35-41,46-69) Bill Sylvester (1970-1984) Bill Lynch (1985-89) Bob Bartolomeo (1990-91) Ken LaRose (1992-2001) Kit Cartwright (2002-05) Jeff Voris (2006- ) Years W 3 8 2 8 1 3 1 2 5 4 2 13 1 1 3 10 1 5 1 4 7 20 1 0 6 37 3 14 2 5 3 10 2 5 32 165 15 84 5 36 2 14 10 46 4 7 4 24 L 0 3 0 1 12 3 0 6 3 3 24 5 14 9 12 11 10 99 65 12 7 54 36 21 T 1 0 0 1 0 1 0 4 0 1 2 1 2 1 0 2 0 13 2 3 1 0 0 0 Pct. .944 .727 1.000 .625 .250 .794 1.000 .600 .625 .563 .457 .083 .717 .604 .294 .478 .333 .619 .563 .735 .659 .460 .163 .533 Butler Year-By-Year Scores 1887 Coach: Clint Howe 45 Purdue 48 Franklin 24 Hanover (3-0-0) 5 8 10 1888 (No Football) 1889 Coach: Clint Howe 32 Hanover 14 Purdue (2-0-0) 0 0 1890 Coach: Clint Howe 0 DePauw 18 DePauw 22 Wabash 12 Purdue (3-0-1) 0 0 6 10 1891 Coach: Unknown 20 DePauw 52 State Univ. 6 Michigan 26 State Univ. 28 Wabash 34 Cincinnati 0 Purdue (4-3-0) 32 6 42 6 6 10 58 1892 Coach: Unknown 18 Earlham 10 Indiana 14 Wabash 12 Dayton 6 Purdue 20 DePauw (5-1-0) 6 6 12 0 40 18 1893 Coach: Unknown 28 Wabash (1-0-0) 24 1894 Coach: Joseph Flint 0 Purdue 38 DePauw 64 Kokomo 12 Danville Normal 34 Rose Poly 58 Wabash 6 Indianapolis Art (6-1-0) 30 6 0 0 0 0 4 1895 Coach: Joseph Flint 6 Wabash 0 Miami 34 Indiana 18 Knightstown (2-2-0) 10 6 2 0 1896 Coach: Unknown (4-3-0) 22 Franklin 6 6 Indiana 22 18 Rose Hulman 0 0 Miami 6 16 Miami 4 13 Earlham 0 0 Indianapolis Ath. Club 14 1897 Coach: James Zink 24 Franklin 16 Miami 4 Earlham (3-0-0) 6 4 0 1898 Coach: T. Williamson 4 Earlham 11 Franklin 0 Earlham 0 DePauw (2-1-1) 0 0 10 0 1899 Coach: Walter Kelly 11 Franklin 0 DePauw 0 Franklin 0 Earlham (1-3-0) 0 28 17 10 1900 Coach: Walter Kelly 0 Franklin 0 Earlham 0 Franklin 10 DePauw (1-3-0) 17 33 37 7 1901 Coach: Walter Kelly 30 Earlham (1-0-0) 0 1902 Coach: Walter Kelly 0 Cincinnati 17 Wabash 0 DePauw 0 Franklin (1-3-0) 6 12 32 5 1903 Coach: Walter Kelly 0 DePauw 0 Wabash 0 Rose Hulman (0-3-0) 18 46 31 1904 Coach: R. Wingard 47 State Normal 32 Miami 56 Danville Normal 17 State Normal 28 Earlham 16 Hanover 0 Wabash (6-1-0) 0 0 5 0 8 6 51 1905 Coach: R. Wingard 101 State Normal 6 Wabash 0 Indiana 31 Winona Tech 47 Shelbyville 6 Rose Hulman 10 Wittenberg 17 Miami 18 DePauw 64 Franklin (7-2-1) 0 0 31 0 0 6 12 0 17 0 1906 Coach: Don Robinson (1-0-0) 17 Irvington Ath. Club 0 1907 Coach: John McKay 0 State Normal 5 Rose Hulman 5 Winona 0 Franklin 6 Earlham 0 Hanover (1-3-2) 0 16 0 0 34 22 1908 Coach: John McKay 22 Winona 18 Hanover 31 Earlham 23 Franklin 10 Hanover 6 Rose Hulman (5-0-1) 5 0 0 0 0 6 1909 Coach: Walter Gipe 25 Winona 18 Franklin 23 Hanover 6 Earlham 6 DePauw 6 Rose Hulman 0 Cincinnati 12 Wabash (5-3-0) 0 0 5 0 12 12 22 0 1910 Coach: John McKay 34 Georgetown 5 Hanover 18 Moores Hill 0 Wabash 0 Indiana 3 DePauw 6 Earlham 0 Miami (4-3-1) 0 3 0 48 33 0 17 0 1911 Coach: Dave Allerdice 19 Franklin 0 Transylvania 9 Notre Dame 22 Cincinnati 6 Earlham 3 DePauw 6 Rose Hulman 45 Moores Hill (4-3-1) 0 0 27 12 39 0 11 0 1912 Coach: G. Cullen Thomas (5-3-0) 54 Hanover 0 25 Franklin 0 0 Wabash 47 13 Earlham 0 52 Moores Hill 14 27 Transylvania 0 3 DePauw 17 6 Rose Hulman 13 1913 Coach: G. Cullen Thomas (2-4-1) 7 Kentucky State 21 10 Wabash 6 14 Franklin 7 0 Earlham 0 0 Louisville 21 0 DePauw 12 19 Rose Hulman 20 1914 Coach: G. Cullen Thomas (4-2-0) 0 Georgetown 13 7 Earlham 6 17 Hanover 16 0 Transylvania 47 7 DePauw 0 6 Franklin 0 1915 Coach: G. Cullen Thomas (1-6-0) 0 Kentucky State 33 16 Franklin 20 0 Rose Hulman 7 7 Wabash 35 0 DePauw 39 20 Hanover 7 19 Earlham 34 1916 Coach: G. Cullen Thomas (3-5-0) 3 Kentucky State 39 83 Meron 0 27 Earlham 0 0 Wabash 56 7 Louisville 19 0 DePauw 20 14 Franklin 39 13 Rose Hulman 7 1917 Coach: G. Cullen Thomas (3-3-0) 0 Kentucky State 33 0 Millikin 34 27 Hanover 7 6 Franklin 0 6 Earlham 0 0 Rose Hulman 25 1918 Coach: G. Cullen Thomas (2-1-1) 32 Hanover 6 6 Franklin 2 0 Miami 52 0 Rose Hulman 0 1919 Coach: Joe Mullane 0 Wabash 0 DePauw 7 Rose Hulman 0 Hanover 0 Earlham 0 Franklin (0-5-1) 67 76 21 0 6 14 1920 Coach: Pat Page 7 Wittenberg 53 Hanover (7-1-0) 20 7 74 13 21 35 9 39 Wilmington Earlham Franklin Rose Hulman Chicago YMCA Georgetown 0 7 10 7 0 0 1921 Coach: Pat Page 19 Denison 70 Rose Hulman 122 Hanover 33 Earlham 0 Wabash 7 Chicago YMCA 3 Michigan Aggies 28 Franklin (6-2-0) 6 6 6 7 14 14 2 0 1922 Coach: Pat Page 6 Wilmington 14 Franklin 16 Chicago YMCA 10 Illinois 57 Earlham 9 Wabash 19 Rose Hulman 19 DePauw 3 Notre Dame 7 Bethany (8-2-0) 0 0 0 7 0 7 0 0 32 29 1923 Coach: Pat Page 39 Hanover 7 Illinois 2 Wabash 13 DePauw 16 Bethany 13 Franklin 7 Notre Dame 19 Haskell 26 Chicago YMCA (7-2-0) 0 21 0 0 0 7 34 13 6 1924 Coach: Pat Page 21 Hanover 10 Franklin 10 Illinois 7 Centenary 12 Wabash 26 DePauw 0 Iowa 0 Ohio Wesleyan 7 Haskell (4-5-0) 6 7 40 9 0 7 7 24 20 1925 Coach: Pat Page 28 Earlham 6 DePauw 13 Illinois 23 Franklin 0 Wabash 38 Rose Hulman 7 Minnesota 10 Dayton 9 Centenary (5-2-2) 0 6 16 0 0 0 33 7 0 1926 Coach: Tony Hinkle 38 Earlham 70 Hanover 7 Illinois 7 Franklin 10 DePauw 0 Lombard 0 Wabash 0 Minnesota 6 Dayton (3-6-0) 0 0 38 0 21 18 13 81 20 1927 Coach: George Potsy Clark (4-3-1) 46 Muncie Normal 12 58 VALPARAISO 0 0 At Illinois 58 7 FRANKLIN 7 25 DePauw 6 6 LOMBARD 19 13 Wabash 6 0 At Michigan State 25 1928 Coach: George Potsy Clark (6-2-0) 0 At Northwestern 14 55 FRANKLIN 0 40 DANVILLE C. NORMAL 0 13 WASHINGTON (MO.) 0 12 BALL STATE 7 0 ILLINOIS 14 24 EARLHAM 0 26 TUFTS 3 1929 Coach: George Potsy Clark (4-4-0) 13 ILLINOIS WESLEYAN 9 0 At Northwestern 13 6 HASKELL 13 6 At New York Univ.# 13 13 At DePauw 0 14 WABASH 0 0 MILLIKIN 6 33 LOYOLA (LA.) 13 #Game played at Yankee Stadium. 1930 Coach: Harry M. Bell (2-7-0) 46 INDIANA CENTRAL* 0 7 OHIO 12 0 Illinois 27 0 ST. LOUIS 7 13 Wabash 7 0 At Loyola (La.) 33 0 Purdue 33 0 Haskell 27 0 Marquette 25 •First night game in the Butler Bowl. 1931 Coach: Harry M. Bell (3-5-0) 6 Franklin 7 0 At Ohio 40 34 BALL STATE 0 61 Louisville 6 2 Dayton 26 13 WABASH 0 0 MARQUETTE 21 7 At Washington (D.C.) 32 1932 Coach: Fred Mackey 13 Ball State 7 Cincinnati 7 Millikin 0 Wabash 14 Franklin 0 Drake 0 Dayton (2-4-1) 12 13 13 34 0 0 7 1933 Coach: Fred Mackey 2 Franklin 19 Ball State 6 Drake 24 Evansville 0 Wabash 7 Cincinnati 7 Valparaiso 12 Washington (2-6-0) 16 2 26 6 12 34 20 36 1934 Coach: Fred Mackey (6-1-1) 13 Ball State 4 25 Franklin 0 50 Danville C. Normal 0 12 Indiana State 0 0 Wabash 0 7 Washington 32 6 Manchester 0 12 Valparaiso 7 1935 Coach: Tony Hinkle 29 Louisville 12 Evansville 71 Hanover 33 Indiana State 39 Valparaiso 20 Wabash 18 Franklin 7 Western State (7-1-0) 0 0 7 7 0 0 0 19 33 Butler Year-By-Year Scores 1936 Coach: Tony Hinkle 40 Evansville 12 Cincinnati 6 Chicago 26 Manchester 9 Wabash 64 Franklin 41 Valparaiso 13 Western State (6-0-2) 0 12 6 0 7 0 0 7 1937 Coach: Tony Hinkle (5-2-1) 7 At Purdue 33 13 At Cincinnati 0 33 VALPARAISO 0 51 EVANSVILLE 0 12 WASH. & JEFFERSON 0 13 At DePauw 0 0 WABASH 0 13 WESTERN STATE 14 1938 Coach: Tony Hinkle (4-4-0) 12 Ball State 6 6 Purdue 21 0 George Washington 26 12 DePauw 0 35 Ohio Wesleyan 0 27 Wabash 0 0 Western State 13 21 Washington (Mo.) 27 1939 Coach: Tony Hinkle (7-0-1) 16 BALL STATE 0 12 OHIO 7 34 INDIANA STATE 0 13 GEORGE WASHINGTON 6 33 At DePauw 0 6 WASHINGTON (MO.) 6 55 WABASH 0 12 At Western State 0 1940 Coach: Tony Hinkle (4-4-1) 27 ST. JOSEPH’S 6 0 At Purdue 28 7 At Ohio 7 6 XAVIER 13 19 At Wabash 12 19 At Washington (Mo.) 27 32 DePAUW 6 25 BALL STATE 0 7 TOLEDO 20 1946 Coach: Tony Hinkle 19 Eastern Illinois 13 Indiana State 0 Western Michigan 41 DePauw 20 Ball State 25 Wabash 31 St. Joseph’s 25 Valparaiso (7-1-0) 12 7 19 6 6 7 6 0 1947 Coach: Tony Hinkle (5-3-1) 6 BALL STATE 6 7 At Ohio 14 21 ST. JOSEPH’S 0 14 At Wabash 0 21 WESTERN MICHIGAN 20 35 DePAUW 0 0 WESTERN RESERVE 6 27 VALPARAISO 6 19 At Cincinnati 26 1948 Coach: Tony Hinkle 68 Indiana Central 14 Evansville 0 Western Reserve 0 Washington (MO) 7 Cincinnati 20 Wabash 7 Western Michigan 6 Ohio (3-5-0) 7 13 6 7 16 7 20 14 1949 Coach: Tony Hinkle 7 Evansville 14 Wabash 6 Western Reserve 47 Indiana State 0 Washington (Mo.) 0 ILLINOIS STATE 6 Western Michigan 0 Ohio (2-6-0) 24 7 28 14 7 14 40 14 1950 Coach: Tony Hinkle (4-4-1) 12 At Evansville 14 7 WABASH 7 14 OHIO 21 33 At Ball State 7 7 MIAMI (OHIO) 42 25 At Western Reserve 14 13 At Western Michigan 34 25 WASHINGTON (MO.) 20 32 At Indiana State 0 1941 Coach: Tony Hinkle 6 ST. JOSEPH’S 7 Xavier 6 Western State 13 Ball State 20 DePauw 7 Ohio 26 Wabash 18 Toledo 40 Washington (MO) (5-4-0) 13 40 14 6 6 20 0 2 13 1942 Coach: Pop Hedden 14 XAVIER 0 Indiana 0 Illinois 0 Ohio 0 WABASH 7 Western Michigan 39 DePAUW 12 TOLEDO 0 ST. JOSEPH’S 1951 Coach: Tony Hinkle (4-4-1) 7 VALPARAISO 41 7 WEST. RESERVE 6 26 At Wabash 26 20 BALL STATE 14 6 At St. Joseph’s 12 27 EVANSVILLE 12 0 WEST. MICHIGAN 20 13 At Washington (Mo.) 20 14 At Indiana State 7 (2-7-0) 21 53 67 6 6 13 0 0 6 1952 Coach: Tony Hinkle (5-3-1) 25 At Evansville 20 47 NORTH CENTRAL 6 25 WABASH 27 28 At Ball State 6 33 ST. JOSEPH’S 0 13 INDIANA STATE 13 13 At Valparaiso 14 33 WASHINGTON (Mo.) 20 14 At Western Reserve 42 1943-44 (No football) 1945 Coach: Pop Hedden 7 Eastern Illinois 56 Earlham 32 Franklin 56 Manchester 2 Ball State 0 Valparaiso 34 (3-3-0) 12 7 6 0 16 6 1953 Coach: Tony Hinkle (6-2-0) 27 EVANSVILLE 0 24 At Wabash 20 25 BALL STATE 7 47 At St. Joseph’s 13 47 INDIANA STATE 12 32 VALPARAISO 20 14 At Washington (Mo.) 27 20 WESTERN RESERVE 21 1954 Coach: Tony Hinkle (4-4-1) 21 At Evansville 14 14 WABASH 21 13 At Ball State 26 40 ST. JOSEPH’S 12 38 INDIANA STATE 26 7 At Valparaiso 39 6 WASHINGTON (MO.) 25 13 INDIANA CENTRAL 7 13 At Western Reserve 13 1955 Coach: Tony Hinkle (3-5-0) 14 EVANSVILLE 45 26 At Indiana State 19 20 BALL STATE 13 13 At St. Joseph’s 28 18 DePAUW 7 14 VALPARAISO 24 12 At Wabash 14 20 At Washington (Mo.) 41 1956 Coach: Tony Hinkle (6-2-0) 34 At Evansville 7 32 INDIANA STATE 0 28 At Ball State 12 6 ST. JOSEPH’S 31 19 At DePauw 13 20 At Valparaiso 6 26 WABASH 7 20 WASHINGTON (MO.) 21 1957 Coach: Tony Hinkle (7-2-0) 0 BRADLEY 13 14 At Wabash 6 13 At St. Joseph’s 34 27 INDIANA STATE 0 27 At Valparaiso 0 27 BALL STATE 7 19 At Evansville 7 26 DePAUW 13 41 WASHINGTON (MO.) 13 1958 Coach: Tony Hinkle (8-1-0) 39 At Bradley 19 40 WABASH 6 6 ST. JOSEPH’S 0 31 At Indiana State 8 34 VALPARAISO 0 7 At Ball State 14 28 EVANSVILLE 14 30 At DePauw 0 20 At Washington (Mo.) 12 1959 Coach: Tony Hinkle (9-0-0) 27 BRADLEY 8 28 At Wabash 8 20 At St. Joseph’s 7 41 INDIANA STATE 6 10 At Valparaiso 7 27 BALL STATE 0 33 At Evansville 14 21 DePAUW 3 48 WASHINGTON (MO.) 13 1960 Coach: Tony Hinkle (8-1-0) 18 At Bradley 12 40 WABASH 7 6 ST. JOSEPH’S 24 20 At Indiana State 13 27 VALPARAISO 20 27 At Ball State 0 34 EVANSVILLE 6 13 At DePauw 6 33 At Washington (Mo.) 6 1961 Coach: Tony Hinkle 34 BRADLEY 48 BALL STATE 34 At Wabash 12 At DePauw 27 ST. JOSEPH’S 26 At Indiana State 14 VALPARAISO (9-0-0) 23 6 7 6 7 0 2 30 26 At Evansville WASHINGTON (MO) 1962 Coach: Tony Hinkle 34 At Bradley 28 At Ball State 14 WABASH 21 DePAUW 0 At St. Joseph’s 41 INDIANA STATE 16 At Valparaiso 41 EVANSVILLE 13 At Marshall 7 7 (5-2-2) 16 28 14 18 6 20 14 0 26 1963 Coach: Tony Hinkle (8-1-0) 13 At Morehead State 31 35 BRADLEY 27 13 BALL STATE 0 26 At Wabash 21 14 At DePauw 12 27 ST. JOSEPH’S 0 7 At Indiana State 6 27 VALPARAISO 12 32 At Evansville 14 1964 Coach: Tony Hinkle (4-4-1) 7 MOREHEAD STATE 26 21 At Bradley 28 14 At Ball State 28 7 WABASH 7 9 DePAUW 6 41 At St. Joseph’s 2 7 INDIANA STATE 2 14 At Valparaiso 23 48 EVANSVILLE 21 1965 Coach: Tony Hinkle (6-3-0) 41 TAYLOR 6 27 At Indiana State 7 21 ST. JOSEPH’S 12 21 At Valparaiso 23 42 EVANSVILLE 0 7 At Ball State 22 14 DePAUW 8 7 At Akron 14 27 WEST. KENTUCKY 20 1966 Coach: Tony Hinkle (4-5-0) 6 Northern Illinois 34 28 Indiana State 6 20 St. Joseph’s 7 12 Valparaiso 15 26 Evansville 7 14 Ball State 17 14 DePauw 7 14 Akron 20 7 Western Kentucky 35 1967 Coach: Tony Hinkle (2-7-0) 7 NO. ILLINOIS 24 7 At Indiana State 23 27 ST. JOSEPH’S 2 7 At Valparaiso 21 7 EVANSVILLE 24 7 At Ball State 65 20 DePAUW 21 14 At Wabash 0 14 WEST. KENTUCKY 36 1968 Coach: Tony Hinkle (2-7-0) 7 AKRON 32 0 At Western Kentucky 35 12 INDIANA STATE 28 49 At St. Joseph’s 14 7 VALPARAISO 10 7 At Evansville 44 21 BALL STATE 24 7 At DePauw 30 26 WABASH 9 1969 Coach: Tony Hinkle (3-6-0) 0 AKRON 52 57 INDIANA CENTRAL 0 7 34 6 17 31 9 38 At Ball State At DePAUW At Wabash At St. Joseph’s INDIANA STATE At Evansville VALPARAISO 36 23 17 20 54 14 20 1970 Coach: Bill Sylvester (3-6-1) 0 At Akron 34 13 BALL STATE 26 14 At DePauw 6 21 WABASH 21 24 ST. JOSEPH’S 26 0 At Indiana State 61 18 EVANSVILLE 31 34 At Valparaiso 31 0 At Western Kentucky 14 35 INDIANA CENTRAL 0 1971 Coach: Bill Sylvester (3-7-0) 0 At Akron 24 0 At Ball State 27 15 DePAUW 13 14 At Wabash 0 6 At St. Joseph’s 24 21 INDIANA STATE 14 8 At Evansville 21 12 VALPARAISO 48 0 WEST. KENTUCKY 31 12 INDIANA CENTRAL 17 1972 Coach: Bill Sylvester (5-5-0) 7 AKRON 34 41 BALL STATE 50 34 At DePauw 7 55 WABASH 6 33 ST. JOSEPH’S 8 21 At Indiana State 49 6 EVANSVILLE 10 17 At Valparaiso 10 6 At Western Kentucky 35 8 At Indiana Central 7 1973 Coach: Bill Sylvester (5-5-0) 19 At Akron 51 14 At Ball State 52 13 At St. Joseph’s 7 13 At Wabash 7 12 VALPARAISO 6 13 INDIANA STATE 41 36 DePAUW 21 34 EVANSVILLE 35 6 WEST. KENTUCKY 48 21 INDIANA CENTRAL 14 1974 Coach: Bill Sylvester (8-2-0) 21 WAYNE STATE 14 0 BALL STATE 45 31 At Valparaiso 15 22 WABASH 17 24 At DePauw 20 29 INDIANA CENTRAL 26 27 At Indiana State 56 39 At Evansville 16 64 At St. Joseph’s 26 35 FRANKLIN 28 1975 Coach: Bill Sylvester 21 At Evansville 20 ROSE HULMAN 37 ST. JOSEPH’S 44 At Indiana Central 39 VALPARAISO 35 At Wabash 17 At Wayne State 14 DePAUW 51 At Franklin 28 ST. NORBERT (9-1-0) 19 12 8 7 8 0 21 7 20 15 1976 Coach: Bill Sylvester 28 EVANSVILLE 34 HILLSDALE 18 At Wittenberg (6-4-0) 31 28 21 Butler Year-By-Year Scores 29 24 18 35 23 35 28 At Valparaiso INDIANA CENTRAL At St. Joseph’s FRANKLIN At DePauw WABASH EASTERN ILLINOIS 1977 Coach: Bill Sylvester 13 DAYTON 7 At Hillsdale 3 WITTENBERG 14 VALPARAISO 11 At Indiana Central 17 ST. JOSEPH’S 7 At Franklin 31 At Eastern Illinois 41 DePAUW 28 At Evansville 49 6 28 7 7 12 27 (5-5-0) 45 20 14 7 30 7 42 13 0 20 1978 Coach: Bill Sylvester (5-5-0) 3 EASTERN ILLINOIS 42 17 HILLSDALE 14 6 At Dayton 31 24 At Valparaiso 20 20 INDIANA CENTRAL 6 17 At St. Joseph’s 21 13 FRANKLIN 14 6 At St. Norbert 20 13 At DePauw 12 9 EVANSVILLE 7 1979 Coach: Bill Sylvester 0 At Eastern Illinois 9 At Hillsdale 0 DAYTON 25 VALPARAISO 13 At Indiana Central 23 ST. JOSEPH’S 14 At Franklin 24 ST. NORBERT 24 DePAUW 21 At Evansville (5-5-0) 38 10 24 24 14 28 10 7 14 10 1980 Coach: Bill Sylvester (5-5-0) 17 HILLSDALE 10 0 At Dayton 29 13 At Valparaiso 25 21 FRANKLIN 26 7 At Georgetown 14 10 At Ashland 0 31 EVANSVILLE 20 14 WITTENBERG 35 7 INDIANA CENTRAL 6 0 At St. Joseph’s 24 1981 Coach: Bill Sylvester 0 At Hillsdale 0 DAYTON 10 At Evansville 16 VALPARAISO 21 At Franklin 34 GEORGETOWN 10 ASHLAND 42 At Wittenberg 31 ST. JOSEPH’S 14 At Indiana Central (3-7-0) 37 27 31 0 25 6 38 14 33 16 1982 Coach: Bill Sylvester (7-3-0) 20 WAYNE STATE 7 20 At Dayton 14 7 WITTENBERG 17 27 At Valparaiso 3 6 FRANKLIN 10 39 At Georgetown 0 6 At Ashland 8 20 EVANSVILLE 9 31 At St. Joseph’s 16 14 INDIANA CENTRAL 7 1983 Coach: Bill Sylvester 19 At Wayne State 20 DAYTON 14 At Wittenberg (9-1-1) 6 3 3 41 31 38 24 21 24 26 VALPARAISO At Franklin GEORGETOWN ASHLAND At Evansville ST. JOSEPH’S At Indiana Central NCAA Division II Playoffs 6 At UC-Davis 35 17 14 9 21 7 6 25 1984 Coach: Bill Sylvester (6-4-0) 6 KENTUCKY STATE 7 20 WITTENBERG 9 0 At Dayton 34 33 GEORGETOWN 7 22 At St. Joseph’s 17 16 At Evansville 13 13 ASHLAND 14 34 At Franklin 27 28 VALPARAISO 27 10 At Indiana Central 20 1985 Coach: Bill Lynch (8-2-0) 24 At Kentucky State 0 23 At Wittenberg 24 27 DAYTON 14 31 At Georgetown 18 31 ST. JOSEPH’S 3 21 EVANSVILLE 20 7 At Ashland 17 39 FRANKLIN 10 26 At Valparaiso 15 18 INDIANA CENTRAL 17 1986 Coach: Bill Lynch (5-5-0) 16 At Dayton 17 28 GRAND VALLEY ST. 30 36 ANDERSON 0 32 At St. Joseph’s 22 28 At Evansville 9 14 ASHLAND 24 20 At Franklin 21 28 VALPARAISO 6 25 At Indianapolis 28 31 FINDLAY 12 1987 Coach: Bill Lynch (8-1-1) 19 At Grand Valley State 24 64 At Anderson 0 15 DAYTON 10 27 At Northeast Missouri 22 14 ST. JOSEPH’S 3 28 EVANSVILLE 28 31 At Ashland 6 49 FRANKLIN 14 29 At Valparaiso 22 35 INDIANAPOLIS 20 1988 Coach: Bill Lynch (8-2-1) 29 FERRIS STATE 13 34 NORTHWOOD 11 10 At Central State 55 34 At Dayton 17 43 ST. JOSEPH’S 20 35 KENTUCKY STATE 14 56 At Valparaiso 0 13 At Indianapolis 13 17 ASHLAND 10 27 At St. Ambrose 0 NCAA Divison II Playoffs 6 At Tennessee-Martin 23 1989 Coach: Bill Lynch (7-2-1) 3 At Ferris State 28 35 At Northwood Institute 21 18 GRAND VALLEY ST. 27 23 DAYTON 23 43 At St. Joseph’s 28 23 At Kentucky State 19 31 VALPARAISO 0 31 INDIANAPOLIS 25 17 At Ashland 6 56 ST. AMBROSE 28 1990 Coach: Bob Bartolomeo (5-5-1) 9 At Northern Michigan 10 17 ST. JOSEPH’S 10 0 At Grand Valley State 35 10 At Dayton 14 9 INDIANAPOLIS 9 16 At Wayne State 7 37 VALPARAISO 0 18 At Ferris State 27 17 ASHLAND 3 27 HILLSDALE 16 7 At Saginaw Valley State 17 1991 Coach: Bob Bartolomeo (9-2-0) 28 NO. MICHIGAN 0 37 At St. Joseph’s 10 33 GRAND VALLEY ST. 0 22 At Indianapolis 3 42 WAYNE STATE 7 22 At Valparaiso 2 6 FERRIS STATE 7 14 At Ashland 12 26 At Hillsdale 20 13 SAGINAW VALLEY 10 NCAA Division II Playoffs 16 At Pittsburg State 26 33 0 EVANSVILLE At St. Joseph’s 31 49 1997 Coach: Ken LaRose (6-4-0) 10 At Howard Payne 9 21 ROBERT MORRIS 26 38 At St. Francis (Pa.) 12 17 CLINCH VALLEY 3 7 At Dayton 42 14 At San Diego 24 17 VALPARAISO 19 38 At Evansville 35 14 DRAKE 13 20 ST. JOSEPH’S 13 1998 Coach: Ken LaRose (4-6) 21 ALBION 44 17 At Morehead State 55 13 ST. FRANCIS (PA.) 0 26 At Wesley (Del.) 24 27 DAYTON 31 29 SAN DIEGO 27 10 At Valparaiso 17 7 At Drake 41 39 At Quincy 0 33 LINDENWOOD (OT) 34 1992 Coach: Ken LaRose (8-2-0) 14 At Northern Michigan 0 33 ST. JOSEPH’S 7 10 At Grand Valley State 21 28 INDIANAPOLIS 6 31 At Wayne State 6 42 VALPARAISO 13 7 At Ferris State 35 24 ASHLAND 21 28 HILLSDALE 17 37 At Saginaw Valley State 0 1999 Coach: Ken LaRose (5-5) 27 At Albion (OT) 20 34 MOREHEAD STATE 56 21 At St. Francis (Pa.) 7 34 WESLEY (DEL.) 19 7 At Dayton 42 20 VALPARAISO 38 14 At San Diego 42 6 DRAKE 53 27 QUINCY 12 27 At Lindenwood 16 1993 Coach: Ken LaRose 19 At Hofstra 24 At Georgetown 28 DRAKE 7 HILLSDALE 10 VALPARAISO 6 At Dayton 27 At San Diego 14 At Evansville 27 UAB 21 At Indianapolis (4-6-0) 20 21 3 29 0 28 28 12 31 34 2000 Coach: Ken LaRose (2-8) 37 ST. FRANCIS (IND.) 56 43 At Morehead State 46 41 ST. FRANCIS (PA.) 7 56 At Georgetown (DC) (OT) 57 26 DAYTON 43 23 At Albion 24 7 At Valparaiso 33 37 SAN DIEGO 38 41 At Drake 62 45 At Quincy 12 1994 Coach: Ken LaRose (7-3-0) 0 HOFSTRA 41 42 At St. Xavier 6 31 GEORGETOWN 21 28 At Wis.-Stevens Point 16 28 At Drake 20 14 At Valparaiso 20 31 DAYTON 24 38 SAN DIEGO 21 49 EVANSVILLE 14 14 At UAB 19 2001 Coach: Ken LaRose (5-5) 20 At St. Francis (Ind.) 41 29 MOREHEAD STATE 27 23 DUQUESNE 33 31 ALBION 28 38 VALPARAISO 21 7 At Dayton 45 19 At San Diego 16 39 DRAKE 41 48 QUINCY 27 24 At Southern Utah 49 1995 Coach: Ken LaRose (2-8-0) 17 HOWARD PAYNE 7 3 At Towson State 34 15 At Millikin 27 0 WIS.-STEVENS PT. 37 8 DRAKE 29 42 VALPARAISO 44 13 At Dayton 49 29 THOMAS MORE 37 14 At Evansville 13 16 At San Diego 37 2002 Coach: Kit Cartwright 54 TIFFIN 0 At Florida Internat’l 43 UW-STEVENS PT. 20 At Morehead State 0 At Dayton 23 AUSTIN PEAY 26 SAN DIEGO 48 DRAKE 52 At Valparaiso 28 At Quincy 1996 Coach: Ken LaRose 3 TOWSON STATE 0 At Robert Morris 42 MILLIKIN 6 At Clinch Valley 7 At Drake 29 At Valparaiso 10 DAYTON 34 SAN DIEGO 2003 Coach: Kit Cartwright (2-9) 6 At Tiffin 42 0 At Duquesne 49 27 DAVIDSON 45 7 At UW-Stevens Point 56 7 At Austin Peay 28 16 ST. FRANCIS (IND.) 47 7 At Drake 24 0 DAYTON 40 (3-7-0) 14 38 7 34 51 50 30 3 (4-6) 31 42 29 53 41 40 35 44 22 37 7 25 17 At San Diego VALPARAISO ST. JOSEPH’S 53 21 13 2004 Coach: Kit Cartwright (1-10) 12 At Albion 13 3 TIFFIN 48 7 At Morehead State 15 14 At Davidson 21 21 AUSTIN PEAY 14 7 At St. Francis (Ind.) 35 6 DRAKE 43 10 At Dayton 49 12 SAN DIEGO 41 26 At Valparaiso 31 0 ST. JOSEPH’S 34 2005 Coach: Kit Cartwright (0-11) 23 ALBION* 28 7 At Tiffin 30 13 At Robert Morris 49 21 At Jacksonville 55 10 MOREHEAD STATE 58 7 At San Diego 49 21 VALPARAISO 34 7 ST. JOSEPH’S 23 10 At Drake 56 7 DAYTON 41 28 MISSOURI-ROLLA 45 *--At Univ. of Indianapolis. 2006 Coach: Jeff Voris (3-8) 10 At Albion 31 30 HANOVER 10 14 ROBERT MORRIS 35 7 JACKSONVILLE 31 23 DAYTON 20 3 At San Diego 56 0 DRAKE 29 32 VALPARAISO 10 20 At Missouri-Rolla 35 7 MOREHEAD STATE 14 10 At Davidson 50 2007 Coach: Jeff Voris (4-7) 42 ALBION 14 44 At Hanover 14 11 ST. JOSEPH’S 8 28 MISSOURI-ROLLA (OT) 21 9 SAN DIEGO 56 19 At Drake 37 37 At Valparaiso 42 3 At Morehead State 24 32 DAVIDSON 44 0 At Dayton 61 16 At Jacksonville 24 2008 Coach: Jeff Voris (6-5) 20 At Albion 6 28 FRANKLIN 31 41 At Missouri S & T 15 21 DRAKE 15 56 At Campbell 7 48 At Valparaiso 21 31 MOREHEAD STATE 21 9 JACKSONVILLE 45 21 DAYTON (OT) 28 17 At San Diego 34 34 At Davidson 46 2009 Coach: Jeff Voris (11-1) 42 ALBION 3 49 At Franklin 19 42 HANOVER 21 28 At Morehead State (OT) 21 25 SAN DIEGO 24 23 VALPARAISO 14 23 At Campbell 16 14 DAVIDSON 7 31 At Dayton 28 7 At Jacksonville 36 20 DRAKE 17 28 CENT. CONN. STATE# 23 #--Gridiron Classic. 35 All-Time Letterwinners A Abel, Chris, 1989, 90 Abplanalp, Pat, 1987, 88 Abts, Henry, 1938, 39, 40 Ackman, Eric, 1992, 93, 94 Adams, Bob, 1963 Adams, Joe, 1952, 53 Adams, Mark, 2000, 01 Adams, Ron, 1960, 62, 63 Adika, Brian, 2006, 07, 08, 09 Agnew, Ralph, 1916 Ahrendts, Dick, 1954, 55 Albea, Mark, 1973, 74, 75 Alcala, Richard, 2004, 05, 06 Alcorn, Chad, 1987, 88, 89 Aldrich, Jeff, 2003, 04 Alenduff, Marty, 1964 Allanson, Mark, 1987, 88, 89 Allegretti, Carl, 1982 Allegretti, Jim, 1991 Allegretti, Paul, 1986 Allen, Edwin, 1928, 29, 30 Anderson, Cory, 2007 Andreadis, Lou, 1994, 95 Andress, Dave, 1972 Annee, Joe, 1985, 86, 87 (C) Annee, Louis, 1957 Anthony, Jim, 1969 Anthony, Leslie, 1903 Arbo, John, Jr., 1998 Armstrong, Scott, 1933, 34 Aronson, Mike, 1968 Atkins, Rick, 1991 Atkins, V. A., 1986, 88 Aton, Ross, 1995 Attaway, Al, 1968, 69, 70 (C) Austin, Gene, 1988, 89, 90, 91 Austin, Steve, 1992, 93 Avington, Ken, 1956, 57, 58 B Babinec, John, 1971, 73 Badger, Everett, 1912 Bailey, David, 1997, 98, 99 Bailey, Van, 1964, 65, 66, 67 Baird, Ron, 1987, 88, 89, 90 Baker, Bill, 1978, 79, 80, 82 Baker, Charles, 1889, 90, 91, 92 Baker, James, 1954, 55 Baker, Stephen, 1928 Baldwin, Chris, 1993 Barclay, Jim, 1976, 77 Barker, Ted, 1967, 68 Barnard, Scott, 1979, 80, 81 (C) Barnes, Chris, 1987 Barnes, George, 1996 Barney, Doug, 1962, 63 Barnhorst, Jay, 1984, 85, 86 Barr, Richard, 1949 Barron, John, 1986, 87, 88 Barthel, Tim, 1983, 84 Bartolomeo, Bob, 1974, 75, 76 Bastian, Bob, 1920 Baszner, Adam, 2002 Batalis, Nick, 1995, 96, 97, 98 Batrich, Donald, 1945 Batts, Roscoe, 1933, 34, 45 Bauer, Brian, 2000 Bauermeister, Carl, 1929, 30 Bays, Shawn, 1991, 92, 93 Baytala, Mike, 1998 Bearden, Cody, 2007 Beatty, Bart, 1996, 97 Beaverson, Neil, 1973, 75, 76 Beck, Art, 1962, 63, 64 Beck, Jeff, 1992 Beimesche, Andy, 1994 Belcher, William, 1936 Belden, Jim, 1962, 63 Belden, Randy, 1968, 69, 70 (C) Bell, Jim, 1981, 82, 83, 84 Belmonte, Efres, 1978, 79, 80 Benbow, Don, 1959, 60, 61 (C) Benjamin, John, 1955 Bennett, Mike, 2005, 06, 07, 08 Bennett, Paul, 1951, 53 Bennett, Richard, 1947, 48 Berglund, Brent, 1990, 91, 92 Berndt, Dick, 1952, 53, 54 Berninger, Mark, 1996 Bertke, Brian, 1982, 83 Bertuglia, Lennie, 1977, 78 Betz, David, 2000 36 Biddle, Herbert, 1962 Bidstrup, Dick, 1948, 49 Billick, Larry, 1978, 80 Biven, James, 1945 Black, Arthur, 1921, 25 Black, Damon, 1993, 94 Blackaby, Inman, 1935, 36, 37 (C) Blanchford, Vince, 1993, 94 Bland, Eugene (Tiny), 1945 Blanks, Derek, 1982, 83 Blare, George, 1940 Blasenak, Jason, 1995, 96, 97 Blessing, Robert, 1922, 23 Blevins, Rob, 2007, 08 Blum, Norm, 1968 Boa, Andy, 1935, 36, 37 Bohnert, Mark, 1972, 74, 75 Bohrer, Dan, 2006, 07, 08, 09 (C) Bole, Randy, 1972, 74, 75 Bolen, Terry, 1991, 92, 93 Bolger, Grove, 1984, 85, 86 Boltin, Charles, 1952, 53, 54 Bonham, Earl, 1916 (C) Booz, Ken, 1930, 31 (C) Bopp, Ed, 1966, 67, 68 (C) Bork, Bill, 1957, 58, 59 Bork, William, 2008, 09 Borrell, Adam, 1993, 94, 95 Bovenkerk, Greg, 1998 Bowers, David, 1995, 96 Bradford, Derek, 2007, 08, 09 Brand, Harry, 1903 Brandt, Ralph, 1930, 32 Branson, DeWayne, 1985, 86 Branyon, Jared, 2006, 07, 08 Braun, Gregg, 2000, 01 Breeden, Landy, 1979, 80, 81, 82 (C) Britt, Kevin, 1979, 81, 82 Brocchi, Mario, 1999, 00, 01 Brock, Bob, 1967, 68, 69 Brock, Ray, 1929, 30 Brocklesby, Ryan, 2001, 02, 03 Broden, Tom, 1942 Broderick, Charles, 1936, 37, 38 (C) Brodine, Jeff, 1964, 65 Brolsma, Kevin, 2003, 04, 05, 07 Brown, Archie, 1916 Brown, Charles, 1936 Brown, John, 1955, 61, 62 Brown, Leonard, 1969, 70, 71 Brown, Mark, 1903 Brown, Pat, 1970 Brown, Paul, 1919, 20, 21 Brown, Phil, 1920 (C), 21, 22 Brown, Robert, 1933, 34, 35 Brown, T. J., 2005, 06, 07 Bruner, Ralph, 1920 Bruton, Kyle, 2001, 02 Buchanan, Bill, 1887 Buchanan, Jerry, 1981, 83 Buchanan, Phil, 2003, 04, 05, 06 Buckner, Lawrence, 2002 Bugg, Bill, 1928 Bunn, Heath, 1994, 95 Bunnell, Kermit, 1932, 33, 34 Bunt, Ron, 1982, 83, 84, 85 (C) Burdette, Cody, 1935, 36 Burgner, Dan, 1964, 65, 66 Burker, John, 1967, 68 Burkett, Kip, 1977, 78 Burkhart, Clarence, 1911, 12 Burks, Jeff, 1992, 93 Burton, Chris, 1980 Burton, Kelly, 1995 Bush, Dave, 1960, 61, 62 Bush, Kyle, 2007 Butcher, Mark, 1975 Butler, Bert, 1960, 61, 62 Butler, Jerry, 1960 Butler, Mike, 1982 Butters, Tom, 1969 C Cahill, Andy, 2002 Caito, Mike, 1969, 70, 71 Caldicott, Nick, 2008, 09 Calvert, Mark, 1977, 78 Cameron, Harry, 1891 Campbell, Don, 1951, 52, 53 Campbell, Justin, 2001, 02, 03 (C), 04 Campbell, Travis, 1995, 96, 97 Canfield, Vincent, 1924 Capizzi, Lou, 2000, 01, 02 (C) Caporale, Egidio, 1958, 59 Caporale, Louis, 1954 Captain, Ron, 1962, 63, 64 Caranddo, Richard, 1966 Carbone, Dean, 1964 Carlberg, Chad, 1996, 97, 98, 99 (C) Carr, Don, 1960 Carson, Andy, 1968, 69 Carter, Ben, 1932 Carter, Glen, 1922 Carwile, John, 1980, 81, 82, 83 Casselman, Bob, 1977 Catanella, Ken, 1968 Cavosie, John, 1928, 29 Cecil, Carl, 1922, 24, 25 Celarek, Frank, 1939 Celarek, Kevin, 1970 Chakos, Ted, 1981, 82 Chaleff, Boris, 1942 Chandler, Scott, 1953, 54, 55 Chapman, Eric, 1982, 83, 84 Chapman, Jerry, 1974, 75 Chapman, Ralph, 1947, 48, 49 Chappuis, Mark, 1973, 74, 75 Chase, Anton, 2000, 01 Chaulk, Joe, 1974, 75, 77 Chelminiak, John, 1948, 49, 50 Chelovich, Colin, 2001, 02, 03, 04 (C) Cheviron, Mike, 1982, 83, 84, 85 Christner, Phil, 2003, 04, 05 (C) Chrobot, Mike, 1975, 76, 77, 78 Chrobot, Rob, 1984 Cilella, Mike, 1976, 77 Clark, David, 1982, 83 Clark, George, 1887 Clark, Jon, 2003, 04 Clarkson, Nick, 2006, 07, 08 Clarkson, Taylor, 2009 Clausen, Frank, 1903 Clayton, Steve, 1970, 71, 72, 73 (C) Cline, Howard, 1967, 68 Coachys, Jim, 1967 Cochran, Mike, 1987, 88, 89, 90 (C) Coddington, Addison, 1934 Coe, John, 1955 Coffman, Chris, 2000 Cohen, Bennie, 1940 Cole, Ryan, 2003 Collier, George, 1925 Collier, Harrison, 1926, 27 (C) Collins, Matt, 2000, 01, 02 (C) Collins, Rob, 1984, 85, 86 Colway, Gene, 1921 Comotto, Nick, 2007, 08, 09 Compton, John, 1931, 32 Compton, Mel, 1903 Compton, Todd, 1998 Condon, Kyle, 1994, 95, 96, 97 Condon, Louis, 1946 Conley, Bob, 1968 Conn, Scott, 2005, 06, 07 Conner, Dave, 2001 Conner, Kyle, 1997, 99, 00 (C) Connor, Bill, 1936, 37, 38 Connor, Bob, 1937, 38, 39 (C) Connor, Bob, 1978, 79, 80 Conrad, Carson, 1931 Constantino, Silvio, 1937, 38 Cook, Brandon, 2002 Cook, Scott, 1980, 81, 82, 83 Cooks, Jeff, 1994 Cooper, Howard, 1946 Cooper, Jerry, 1978, 79 Cooper, Ron D., 1970 Cooper, Ron L., 1970, 71, 72 Corbett, Mark, 1970, 71, 72, 73 Cornelius, Ed, 1946 Cornelius, George, 1916 Cornelius, Pem, 1946, 49, 50 Cornell, Andrew, 1992 Cosgrove, Walter, 1932 Cosler, Rob, 2008, 09 Costas, Spero, 1934, 35, 36 (C) Cottrell, Andrew, 2009 Coughlin, Kevin, 1971 Cougill, Brad, 1989, 90, 91 Cousino, Case, 2006 Crable, Brian, 2007, 08, 09 Craney, Jason, 1996, 97, 98, 00 (C) Craver, Jim, 1966, 67, 68 Crawford, Bob, 1964, 65 Crawford, George, 1935, 36 Crawford, John, 1936, 37, 38 Crawford, Stanley, 1938, 39, 40 Crawforth, Tim, 1949, 50 (C) Creach, Bill, 1996 Credille, Kevin, 2009 Crockett, Chuck, 1978, 79, 80, 81 Crosley, Howard, 1928, 29 Cross, James, 1890 Cross, Tom, 1950 Crouch, Dan, 1968, 69 Crumley, Jim, 1950, 51 Cruse, Glen, 1908 Cullum, Carl, 1890, 91 Cunningham, Dan, 1963 Cunningham, Dave, 1974, 75 Cunningham, Eric, 2004, 05 Cunnings, Robert, 1945 Currey, Doug, 1977, 78, 79, 80 Curtis, Richard, 1934, 35 Curto, Stephen, 2003 Cutter, Rob, 1985, 86, 87 (C) D Dainton, Ken, 1971 Daniels, Fred, 1916 Dauch, Andy, 2008, 09 Daugherty, Mike, 1975, 76, 77, 78 Davidson, Frank, 1889, 90 Davidson, Mike, 1982, 83, 84 Davies, Thomas, 1928 Davis, Alex, 2005, 06, 07 Davis, Chester, 1916 Davis, Chris, 2006, 07 Davis, Fred, 1951, 52, 53 (C) Davis, Gordon, 1928 Davis, John, 1952 Davis, Kent, 1999, 00, 01 Day, Bob, 1958, 59, 60 (C) Day, Derek, 2005, 2006, 07, 08 Dayton, Charles, 1931 DeCraene, Don, 1990, 91, 92 DeJaegher, Joe, 1992, 93, 94 Delaney, Dave, 1973 Del Busto, Mike, 1980, 81, 82 DeLuna, Ruben, 1989, 90, 91, 92 (C) Delph, Travis, 2001, 02 Dennison, Chuck, 1964, 65, 68 Depositar, Steven, 2009 Derickson, Kyle, 1999, 00, 01, 02 (C) DeShone, Scott, 1988, 89 DeTrude, Keith, 1973, 74, 75 DeWald, Steve, 1942 Dezelan, Joseph, 1938, 39, 40 (C) Dezelan, Joseph, Jr., 1964, 65 Dick, Andy, 1974 Diggins, Ryan, 2003, 05, 06 Dillon, Husani, 1995, 96, 97 Dimancheff, Boris, 1941, 42 Dinn, George, 1973, 74, 76 Dinwiddie, Will, 1995, 96 Dirksing, Chris, 1990 Disney, James, 1966 Disser, Kelly, 2000, 02, 03 Dixon, Travis, 1992, 93, 94 Dixon, Jim, 1987, 88, 89 Dobkins, Knute, 1942, 46, 47, 48 Doctor, John, 1980, 81, 82, 83 Dodds, Ronald, 1945 Dodson, Roger, 1973, 74, 75 Dold, Leslie, 1942, 46 Dombart, Tadd, 2007, 08, 09 Donovan, Brian, 1994 Donhauser, Dick, 1969 Donner, Chase, 2006, 07, 08 Doss, Bill, 1984, 85, 86 Douglas, Jim, 1957, 58, 59 d’Ouville, Paul, 1987, 88 Dovey, Parmalee, 1934 Dowd, Joe, 1973, 74, 75 Dowdy, D’Andre, 2005 Downham, Bob, 1963, 64 Doyle, William, 1941 Dugger, Doyle, 1939, 40 Dukes, Scott, 1941 Dullaghan, Dan, 1968, 69 Dullaghan, Dick, 1963, 64, 65 Duncan, Don, 1953 Durig, Chris, 2002, 03, 04, 05 Dury, George, 1984, 85, 86 Duttenhaver, Harry, 1920, 21, 22 (C) Dykhuizen, Joe, 1986 E Eagan, Bernard, 1952 All-Time Letterwinners Eaton, Joe, 1985, 86, 87 Eckerle, Mark, 1969, 70 Eckert, Kevin, 1998 Eddie, Mike, 2005 Egbers, Dan, 1979, 80, 81 Eichholtz, Bob, 1951, 52, 53, 54 (C) Eldridge, Gary, 1988, 89, 90, 91 Ellenberger, Norm, 1951, 52, 53 (C) Ellis, Jeff, 1981 Elser, Earl, 1930, 32 Elson, Brian, 1987, 88 Elson, Dave, 1991, 92, 93 Ely, Jeremy, 1995 Emmert, Dan, 2002, 03 England, George, 1951 Ennis, Willard, 1930, 32 Enrico, Jim, 1976, 77 Enright, Dave, 1963, 64, 65 Enright, Kevin, 1988, 89, 90 Ent, Harry, 1942 Eppard, John, 1980, 81 Epperson, Stan, 1968, 69, 70 Erickson, Carl, 1999, 00, 01, 02 (C) Errett, Zach, 1997, 98 Erschen, Marty, 1993, 94 Ertel, Mike, 1988, 89 Esary, Les, 1947, 48 Espich, Robert, 1985, 86, 87 Ewald, Clarence, 1953 Ewing, DeWayne, 1998, 99, 00, 01 (C) Evans, Jon, 1988 Eynotten, Bob, 1932, 33 F Fagan, Mark, 1981, 82 Fairchild, Larry, 1964, 65, 66 (C) Farmer, Bob, 2003, 04, 05 Farmer, Harry, 1947, 48, 49 Fattore, Len, 1961 Feichter, Harold, 1940 Ferree, John, 1922 Ferrell, John, 1914, 16 Fickert, Steve, 1970, 71, 72 (C) Fields, Randy, 1976, 77, 78 Fields, Tom, 1921 Fike, Edward, 1947, 49 Fillenwarth, Jack, 1985, 86, 87 Fine, Marion, 1945 Finney, Charles, 1996, 97 Fischer, Tom, 1973, 74 Fish, Guy, 1950, 51 (C) Fisher, Fred, 1947, 48, 49, 50 Fitzsimmons, Phillip, 1970, 71 (C) Flanigan, Tim, 1978, 79, 80 Fleck, Leslie, 1916 Fleming, Doug, 1985, 86 Fletcher, Bill, 1996 Fletcher, Francis, 1923, 24, 25 Fletcher, Nick, 1997, 99 Flick, Paul, 1995 Florence, Rich, 1962, 63 Flowers, Chris, 1990 Flowers, Dave, 1957, 58, 59 Flowers, Jim, 1989, 90 Floyd, Walter, 1923 Flynn, Mike, 1979 Fodor, Carl, 1955 Foor, Matthew, 2009 Ford, Bruce, 1974, 75, 76, 77 Ford, Jamie, 1994, 95, 96, 97 Forsythe, Chester, 1903 Fouty, John, 1950, 51 Frasca, Joe, 2003, 04, 05 Freas, Tom, 1972, 73 Frendenberger, George, 1928, 29 Freeman, Ken, 1960, 61 Freeman, Phil, 1894, 95 Freeman, Vince, 1985, 86 Freuchtenicht, Dick, 1939, 40 Freyn, George, 1953 (C) Friend, Jason, 2004 Fritz, Jacob, 2007, 08, 09 Fromuth, Alan, 1928 Fuhs, Dan, 1978, 79, 80 Fulaytar, Don, 1960 Fulk, Ralph, 1947 Fuller, Tom, 1990, 91, 92 Furnish, Paul, 1953, 56, 57, 58 (C) Furrey, Bryan, 1995, 96 Fus, Mike, 1985, 86 G Gallagher, Dan, 1959, 61 Galloy, Ryan, 1996, 98 Gamblin, Bill, 1954, 55, 56, 59 Gard, Marty, 1999 Garrett, Harvey, 1925 Garvey, Pat, 1967 Garwood, James, 1938, 39, 40 Gates, Damon, 1966, 67 Gatlin, Dan, 1987, 88, 89 Gatto, Joseph, 1945 Gauer, Greg, 1990, 91, 92 Gegner, Mike, 1983, 84 Geiman, Ken, 1942, 46 Geisert, Herman, 1928 (C) Gerlach, Les, 1951, 52, 53, 54 Giacomantonio, Mark, 2008, 09 Gilbert, Larry, 1966, 67, 68 Gillespie, Jim, 1967 Gillum, Joe, 1987, 88 Gillum, Mike, 1989 Gilmore, Donnie, 2007, 08, 09 Gilson, James, 1940, 41, 42 (C) Gilson, John, 1951, 54, 55, 56 Ginn, Bill, 1974, 75, 76, 77 Ginn, Dave, 1981, 82, 83, 84 (C) Glasgow, Joe, 1998 Glunt, Warren, 1928 Goad, Ryan, 1997, 98, 99 Goens, Larry, 1959, 60, 61 Goens, Mike, 1982, 83, 84 Goettler, Logan, 1998 Goettler, Steve, 1997 Goletz, Mike, 1996, 97, 98, 99 (C) Gollner, Bob, 1951 Golomb, Larry, 1965 Gongwer, Ed, 1887 Gonzales, Dax, 1988, 89, 90, 91 Gooch, Thomas, 1945 Good, Charles, 1912 Good, Noel, 1945 Goshert, Rob, 1975, 76, 77 Graham, Alva, 1920, 21 Gray, Rich, 1969, 70 Gray, Robert, 1962 Gray, Scott, 2007, 08, 09 Green, Gary, 1959, 60, 61 Green, Paul, 1934 Greene, Carlton, 1964 Greenwald, Jake, 2005, 06 Greer, Kent, 1990, 91, 92 Greer, Shane, 1992, 93 Greisl, Kevin, 1977 Grenda, Bob, 1971, 72, 73, 74 (C) Gribbins, Kevin, 1991, 93, 94 Grider, Chris, 2006, 07 Griggs, Haldane, 1921, 22, 24 Grimes, Chris, 1999 Grimes, Ricardo, 1976, 77, 78 Grimm, Lee, 1961, 62, 63 (C) Grissom, Joe, 1951, 57, 58, 59 Griswold, Ron, 1994, 95, 96, 97 (C) Gross, Jim, 1977 Gruber, John, 1972, 73 Guggenberger, Derek, 2006, 07, 08, 09 Guyer, Rick, 1947 H Hacker, Dillon, 1903 Haggard, Gordon, 1928 Hailey, Artis, 2009 Haley, Jr., Joe, 2000, 01, 02, 03 Hall, Archie, 1889 Hall, Bob, 1889, 90 Hall, Richard, 1921 Hall, Steve, 1997, 98, 99 Hall, Tom, 1889, 90, 91 Hallam, Ron, 1950, 51 Halloran, Dan, 1972 Hamilton, Bob, 1942, 46, 47, 48 Hamman, Bruce, 1950, 51 Hanson, Tyler, 2001, 02 Hardee, Craig, 1986, 87, 88 Harding, Thomas, 1937, 38, 39 Harkin, Jeremy, 1995, 96, 97, 98 (C) Harrell, John, 1955, 56, 57 (C) Harrington, Paul, 1976, 77, 78, 79 (C) Harris, Bill, 1973, 74 Harris, Justin, 2001, 02, 03 Harrison, Mike, 1965, 66, 67 (C) Hart, Jim, 2002, 03, 04 (C) Hart, Scott, 1995 Hartley, Mike, 1985, 86, 87 Hartley, Tim, 1908, 09 Hartman, Ray, 1908 Harvey, Stuart, 2008 Haste, Mark, 1985 Hatzikostantis, Alex, 1998, 99, 00 Haugh, Otis, 1903 Hauss, Craig, 1965, 66 Hauser, Craig, 1994, 95 Hauss, Jim, 1936, 37 Hauss, John, 1977, 78 Hayes, James, 1938 Haynes, Chris, 2003, 04, 05 (C) Hazelwood, Nathan, 1999, 00 Heacox, Richard, 1945 Healey, Brian, 1996 Healy, Brendan, 1998, 99, 00, 01 Heard, Sidney, 2003 Hearne, Brian, 1980 Heberet, Fred, 1971 Heckman, Drew, 2002, 04 Hedden, Frank, 1929 Hegwood, Mike, 1984, 85, 86, 87 Hein, Dave, 1972 Heines, Tom, 1955 Heintzman, Rob, 1984 Heise, Jonathan, 2003, 04 (C) Helmerich, Matt, 1998, 99 Helms, Larry, 1959, 60, 61 Helton, Carter, 1922, 23, 24, 25 Hensel, Hiram, 1922, 23, 24, 25 Hess, Ben, 2003, 04 Hess, Dick, 1955 Hess, Jim, 2004, 05, 07 Hiday, Mike, 2005, 06, 07 Hildebrand, Keith, 2005, 06 Hill, Erick, 1990, 91 Hill, Jon, 1991, 92, 93 Hilligrass, Frank, 1916 Hillring, Oscar, 1940, 41 Himmel, Keith, 1970, 71, 72, 73 (C) Hinchman, Hubert, 1928, 29, 30 Hinkle, Don, 1947, 48, 49 Hitch, Ralph, 1924, 25 Hitchcock, Randy, 1979, 80, 81 Hitchcock, Ryan, 2008, 09 Hobson, Rob, 2009 Hockett, Dave, 1962, 63 Hoffman, Jason, 1994, 95 Hoffman, Mark, 1969, 70, 71 Holliday, Keen, 1894, 95 Hollingsworth, Joe, 1947 Hollstegge, Dan, 1984 Holman, Rob, 1983, 84 Holmes, Larry, 1934, 36 Holok, Al, 1967, 68, 69 Holsclaw, Jim, 1995, 96, 97 Holst, Dick, 1961 Holthaus, Brian, 2003, 04, 05, 06 Holthaus, Matt, 1998, 99 Holthaus, Scott, 2000, 01, 02 Hoover, John, 1986 Horvath, Bill, 1942 Horvath, Steve, 1947 Hosier, Maurice, 1928, 29 (C) Hoskins, Jim, 1982, 83, 84 Hosmer, Jordan, 2006, 07 Housand, Kevin, 2004, 05 Howard, Andy, 1979, 80, 81, 82 Howard, Bill, 1941, 42 Huber, Dan, 1990 Huber, Joe, 1992 Huck, Andrew, 2009 Huddleston, Lew, 1887 Huff, Todd, 1996, 97, 98, 99 Huffman, Harold, 1970, 71 Hughes, Wayne, 1931 Hughett, Bill, 1953 Hummell, Frank, 1889 Humphrey, Pete, 1978, 79, 80, 81 Hungate, Harold, 1920, 21, 22, 23 (C) Hunt, Rick, 1992 Hunter, Grant, 2008, 09 Hunter, Ken, 1980 Hurrle, Bob, 1950, 55 Hurrle, Ott, 1946, 47 (C), 48 Hurrle, Otto, 1972, 73 Hurley, Marshall, 1954 Hutson, Mike, 1988, 90, 91 Hutt, Matt, 1991 Hysong, James, 1966, 67 (C) I Inman, Dave, 2003, 04, 05 (C), 06 Iozzo, Tom, 1966 Isenbarger, Tom, 1972, 73 Isom, Justin, 1999 J Jackson, John, 1960, 61 Jackson, Tim, 1964 Jacobs, Jay, 1952 Jacobs, Toby, 1993, 94, 95 (C) Jarrett, Robert, 1929 Jasinski, Mike, 2004, 05 Jefferson, Bill, 1974 Jenkins, John, 1985, 86, 87, 88 Jennings, Dale, 1999, 2002 Jennings, Floyd, 1950 Jennings, Tom, 1989 Jensen, Phil, 1984, 85, 86 Jeter, Mel, 1962, 63 Johnson, Ben, 1950, 53 Johnson, Brandon, 1998, 99 Johnson, Charles, 1951, 52 (C) Johnson, E. Paul, 1972, 73 Johnson, Kevin, 1990, 91 Johnson, Luke, 2007, 08 Johnson, Richard, 1991, 92, 93 Johnson, Spurgeon, 1933 Johnson, Walter, 1930 Johnston, Greg, 1986 Johnston, John, 1965, 66 Joiner, Doug, 1995 Jones, Ben, 2001 Jones, Gary, 1962 Jones, Larry, 1947 Jones, Richard, 1956 Jones, Todd, 1985, 86, 87 Jones, Tom, 1963 Jones, Walter, 1903 K Kalm, Philip, 1945 Kammer, Don, 1941 Kantor, Jerry, 1956 Kappen, Steve, 1985, 86 Kathman, Dave, 1991, 92, 93 Katris, Pete, 1977, 78, 79 Kavanaugh, Jim, 1946 Kavanaugh, Joe, 1946 Kazmierczak, Paul, 1977, 78, 79, 80 Keach, Bob, 1922, 23, 24, 25 Kealing, Marshall, 1931, 32 Keckler, Al, 1958 Keeling, Cliff, 1903 Kehrer, Dick, 1967 Kelleher, Joe, 2003, 04 Keller, Todd, 1976, 77, 78, 79 (C) Keller, Tyler, 2005, 06 Kelly, Don, 1951, 52 Kelly, Joe, 1957, 58 Kelly, Pat, 1993, 94 Kelly, Ryan, 1998, 99, 00 Keltner, Ken, 1962 Kennedy, Mannert, 1953, 54, 56 Kenney, Howard, 1972 Kerbox, Will, 1946 Kerins, Craig, 1999, 00, 01, 02 Kiel, Kip, 1984 Kilgore, David, 1924, 25 Kilgore, Charles, 1932 Kilgore, Marc, 1975 Kimble, Kevin, 1990, 91, 92 (C) Kincaid, Jeff, 2005, 06, 07 King, Pat, 1977 King, Robert, 1945 King, Ron, 1950 Kingsburty, John, 1903 Kiolbassa, Ron, 1988, 89, 90 Kirk, Mike, 1976, 77, 78 Kirk, Pat, 1974, 76, 77 Kirkhoff, Louis, 1912 Kirn, Art, 1969, 70 Kirschner, Arnold, 1968, 69, 70 Kiser, Dwight, 1920, 32 Kiser, William, 1919, 20, 21 (C), 23 Kisselman, Harry, 1967 Kistler, Matt, 1991, 92, 93 Klawitter, Gordon, 1966 Klett, Mike, 1992, 93, 94 Kline, Frank, 1935 Klug, Greg, 1980 Klusman, Tom, 1984, 85, 86, 87 Knapik, Rob, 2002, 03 Knieper, Steve, 1982, 83, 84 Knight, John, 1995, 96 Knock, Reginald, 1931 Knox, James, 1954 Kobli, Matt, 2007, 08 Koblinski, Kyle, 2006 37 All-Time Letterwinners Koch, Jim, 1977, 78 Kocur, Carl, 1988, 89, 90 Kodba, Joe, 1942 Koehnen, Joe, 1987, 88, 89 Koehler, Kurt, 1976, 77 Koenig, Russ, 1975, 76 Kokinda, Jack, 1967 Kokonas, Eric, 2006 Kolkmeyer, Tim, 1979, 80, 81 Kollias, Steve, 1982, 83, 84 Kollins, John, 1957, 58, 59 Konold, David, 1921, 22, 24, 25 Koopman, Jordan, 2008, 09 Kosior, Casey, 1977, 78, 90 Koss, Harry, 1933 Koteff, Robert, 2009 Kotulic, Wayne, 1967 Kozlowski, Ron, 1964, 65, 66 Krause, Frank, 1962, 63 Kreag, Bill, 1937, 38, 39 Krebs, Jack, 1961 Kremer, David, Jr., 1995, 96, 97, 98 Kress, Pat, 1967, 68 (C) Kroger, Bob, 1985, 86 Kruse, William, 1941 Kubal, James, 1937, 38 Kuhn, Chris, 1998, 99 Kuntz, Bill, Jr., 1973, 74 Kuntz, Bill, Sr., 1946, 47, 48, 49 (C) Kuntz, Joe, 1989 Kutschke, Jim, 1965 Kuykendall, Bob, 1950 Kuzmic, Gene, 1953 Kwiatkowski, Mark, 2005, 06, 07, 08 Kyvik, Curtis, 1946, 47, 48, 49 L Lafferty, Adam, 2000, 01, 02 Lamar, William, 2009 Lambert, Bob, 1977, 78, 79 Lanahan, Victor, 1937, 38, 39 Landry, Greg, 1974, 75 Landry, Virgil, 1950, 51 Lane, C. D., 1903 Lang, David, 2008, 09 Langston, Ken, 1992 LaRose, Ken, 1976, 77, 78, 79 (C) Larsen, Jeff, 2009 Larson, Todd, 1988, 89, 90, 91 LaVine, Dave, 1942, 46 Lawson, Rob, 2005 Lawyer, John, 1950, 53 Laymon, Clarence, 1933, 34 (C), 35 (C) Laymon, Dan, 1891 League, Vincent, 1966, 67 Leamon, Charles, 1948 Lee, Mike, 1983, 84, 85 (C), 86 (C) Lees, Ed, 1983, 84 Leffler, Jim, 1968, 69 Leffler, K. C., 1990 Leffler, Ken, 1963, 64, 65 Lehane, Dan, 1954, 55 Leininger, John, 2000, 01, 02, 03 (C) Leonard, Bob, 2001, 02, 03 (C) Leonard, Dennis, 1972, 73, 74 Leslie, John, 1920, 21, 22 Lewis, Edwin, 1911, 12 (C) Lewis, Joe, 1964 Lewis, Richard, 1967, 69, 70 Ligda, Bob, 1974, 75, 76 Lill, Jim, 1967, 68, 69 Lister, John, 1894 (C), 95 (C) Livingston, Rob, 1993, 94 Livorsi, Mike, 1949, 50 Lockhart, Arthur, 1912 Loeffler, Greg, 1981, 82 Logan, Greg, 1968, 69 Logan, Mike, 1984, 85 Logsdon, Tim, 1975, 76, 77 London, Ralph, 1951, 52, 53, 54 (C) Long, Karl, 1957, 58, 59 Long, Phil, 1959, 60, 61 Loop, C., 1898 Loop, Marvin, 1894, 95, 98 Lopez, Davey, 2004 Lord, Jack, 1962, 63 Lumpkin, Tijuan, 1994 Lupus, Peter, 1951 Lutkewitte, Mike, 2005 Lynch, Bill, 1972, 74, 75, 76 (C) Lynch, Jim, 1962, 63 Lyon, Jim, 1966 Lyons, Greg, 1991, 92 38 Lyster, Kevin, 1996, 97, 98 Lytle, Tim, 1999, 00, 01, 02 M Macek, Joseph, 1935, 37 Macharaschwili, Dave, 1988, 89, 90 Maglish, Joe, 1981, 82, 83 Magnuson, Bob, 1960 Maheras, Tom, 1987, 88 Maki, Wilho, 1928 Mahoney, Leo, 1952, 53, 54, 55 Mallonee, John, 1975, 76 Malone, Zach, 1995 Mangeot, Dan, 1990, 91, 92, 93 Mangin, Gene, 1951, 52, 53, 54 Manka, John, 1950, 51 Mann, Henry, 1889 (C), 90 (C) Mann, Russ, 2000, 01, 02 (C) Mariacher, Greg, 1987, 88, 89 (C) Marienthal, Matt, 2001, 02 Marion, Cecil, 1933 Marmion, Mike, 1951, 52 Marrs, Joel, 1999 Marrs, Lucas, 1995, 96, 97, 98 Martin, Brandon, 2000, 01, 02, 03 (C) Martin, Gary, 1977 Martin, Luther, 1934, 35 Martin, Vic, 1996, 97 Marzotto, Chris, 2004, 05, 06, 07 (C) Marzotto, Mike, 2005, 06, 07 Masarachia, Vincent, 1935, 36, 37 Massey, Jeff, 1987, 89 Masters, Nolan, 1954, 55, 56 Maternowski, Charles, 1947, 48, 49 Mattingly, Dan, 1979, 80, 81 Mattingly, Pete, 2008, 09 Maxey, Bob, 1955, 56 McCalip, Robert, 1941, 42 McCanna, Matt, 2003, 04, 05 McCarthy, William, 1928, 29 McCauley, Ed, 1960 McClafin, Bill, 1920, 21 McClarnon, Kevin, 1972, 73 McClelland, Harry, 1935 McClure, Brian, 2005, 06 McConnell, Kurt, 1985, 86, 87 McCool, Chris, 1988 McCowan, George, 1973, 74 McCray, Jim, 1979, 80, 81 McCullough, Joe, 1994 McDaniel, Cameron, 1992, 93, 94 (C) McDaniels, Charles, 1935, 36 McDevitt, Kevin, 1973, 74, 75, 76 McDevitt, Mike, 1970, 71, 72 McDowell, Charles, 1938 McElderry, Tim, 1986, 87 McGann, Colin, 2003, 04, 05, 06 McGary, Chris, 1979, 80, 81, 82 McGeorge, Mike, 1978, 79, 80 McGinley, Mike, 1961, 62, 63 McHale, Eddie, 2007, 08, 09 McHugh, Jeke, 1946 McInerney, Will, 2005, 06 McIntire, Jim, 1958 McKay, Bob, 1908 (C), 1909 McKenna, Jack, 2008, 09 McKenzie, Marshall, 1957 McLinn, Jim, 1946 McMahon, Dave, 2002, 03, 04, 06 (C) McManamon, Eugene, 1931 McNerney, Chester, 1933 McSemek, Ray, 1946, 47, 48, 49 Means, Adell, 1993, 94 Mecum, Ralph, 1930 Medford, Andrew, 2005, 06, 07 Meeker, Ray, 1887 (C), 89, 90 Meier, Frank, 1955 Melloh, Nick, 1991 Meloy, Jim, 1980 Melzoni, Rusty, 1985, 86, 87 Mench, Tom, 1972, 73 Menely, Ron, 1989 Mercer, Phil, 1957, 61 Merlina, Dino, 1981, 82, 83, 84 Merrill, William, 1935, 36, 37 Metrick, Mike, 1995 Metzelaars, Charles, 1940 Metzinger, Dave, 1973 Mewborn, Mike, 1985, 86 Meyers, Claude, 1903 Meyers, David, 2005, 06 Michelakis, Mike, 1992, 93, 94 Michielutti, Eric, 1999, 00, 02 Mickens, Arnold, 1994, 95 (C) Middlekauff, Lance, 1961 Middlesworth, Hugh, 1920, 21, 22, 23 Miedema, Ernie, 1951 Mike, David, 1975 Miles, Joey, 1995, 96, 97 Miller, Andy, 1991, 92 Miller, Doug, 1996, 97 Miller, George, 1889, 90, 91 Miller, Harold, 1941, 42 Miller, James, 1945 Miller, Merle, 1925 Miller, Nate, 2004, 05, 06, 07 Miller, Ray, 1930 Mills, Tom, 1984 Minczeski, Matt, 1976, 77 Minnick, Kelly, 1980, 81, 82, 83 Mitchell, James, 1942 Mitchell, Steve, 1976 Mitschelem, Lyle, 1962, 63, 64 (C) Moan, Steve, 1991 Moore, Bill, 1967, 68 Moore, Damon, 1994 Moore, Jeremy, 1996, 97 Moore, John, 1887 Moore, Ken, 1920 Moore, Paul, 1932, 33, 34 Moore, R. E., 1887 Moore, Ralph, 1931, 32 Morales, Seth, 1998 Morelli, Mark, 1973, 74, 75 Morgan, John, 1911, 12 Morgan, Mike, 2007 Moriarity, Francis, 1942, 46, 47, 48 (C) Morris, Kevin, 1992, 93, 94 Morrison, Ino, 1887 Moseley, Keith, 1984, 85, 86 Moses, John, 1956, 57, 58 Mossey, Harold, 1939, 40 Mozingo, Ernest, 1931 Muckerheide, Lynn, 1968, 69, 70 Mueller, Derk, 1997, 98 Mulholland, George, 1923, 24, 25 Mullis, Don, 1997, 98, 99 Mullane, Dan, 1911, 12 Mullane, J., 1911 Mullane, Price, 1916 Mulvihill, Tom, 2006, 07 Murphy, Chris, 1995, 96, 98 Murphy, John, 1946, 47, 48, 49 Murphy, Kevin, 1984 Murphy, Mike, 1981, 82, 83 Murphy, Scott, 2004, 05 Murphy, Thorn, 2008, 09 Murray, Gene, 1993 Murzyn, Dale, 2004, 05, 06, 07 Muse, Frank, 1887, 89 Musgrave, Emerson, 1934, 35, 36 Muta, Harry, 1973, 74, 75 Myatt, Gene, 1972, 73 Myers, Steve, 2002, 03 N Nackenhorst, John, 1935, 37 Nardini, Billy, 2003, 04, 05 (C), 06 Nardo, Nicholas, 1954, 55, 56, 57 Natfzger, George, 1928, 29 Naylor, Mickey, 1981, 82, 84, 85 Neeme, Emil, 1942 Nelson, Andy, 2002, 03, 04 (C) Nelson, Ian, 2002, 03, 04 Neumeier, Eric, 2002, 03, 04, 05 Newcomer, Dave, 1980, 81, 82 (C) Newell, Rick, 1973, 74 Ney, Bill, 1957 Nichols, John, 1889, 90, 91 Nicholson, Ken, 1953, 54, 55, 56 Niemeyer, John, 1966, 67, 68 Nipper, Bob, 1922, 23, 24, 25 Nixon, Ken, 1945 Noel, Rob, 2005, 06, 07, 08 Nolan, Dan, 1968, 69, 70 Norkus, Bill, 1952 Norman, Clyde, 1937, 38 Norris, Belmont, 1931 Norris, Elwood, 1940, 41 Norris, Paul, 1972, 73 North, Derek, 2004, 05, 06, 07 Northam, John, 1922, 23, 25 Notestine, Eric, 2005, 06, 07, 08 Novack, Mitch, 2002 Nulf, George, 1928 Nyers, Jim, 1951 O O’Banion, Elmer, 1958, 59, 60 O’Banion, Tim, 1973, 74 Ober, Nick, 1998, 99, 00, 01 (C) Oberheiman, John, 1962 Oberting, Dave, 1961 O’Brien, Tom, 1952 Ochs, Kyle, 1990, 91, 92 (C) O’Connor, Charles, 1935, 37 O’Connor, Ed, 1934, 35, 36 O’Connor, Frank, 2003, 04, 05 Offerle, Mike, 1965, 66 Oilar, Cliff, 1957, 58, 59 O’Leary, Tim, 1974, 75, 76 Olinger, Scott, 1982, 83, 84 (C) Olinghouse, Dave, 1954 Oliphant, Frank, 1942 Oliver, Dave, 1973, 74, 75 Olszewski, Bob, 2009 Olszta, Rick, 1999, 00 Opatkiewicz, Mark, 1976, 77 Opel, Doug, 1978, 79, 80 Oppenlander, Ben, 1971, 74 Orban, Chuck, 1988, 89, 90 (C) O’Rourke, Jason, 1992, 93, 94 Orphey, Steve, 1966, 67, 68 (C) Osborne, Pat, 1894, 95 P Pachacz, Greg, 2004, 05, 06, 07 (C) Page, Paul, 1984, 85, 86 Paliska, Steve, 2004 Palmer, Jeff, 1983, 84, 85 Parker, Ben, 1995, 96, 97 Parker, Ed, 1894, 95 Parks, Noble, 1980, 81, 82, 83 Paul, Gordon, 1922, 23, 24, 25 Paul, Judson, 1928 Paul, Justus, 1911, 12 Paulson, Craig, 1974, 75 Pavey, Jesse, 1912 Peconge, Mike, 1979, 80, 81, 82 Pedigo, Bob, 1956, 57 Peebles, Julian, 1967 Pence, Tony, 1978, 79, 80, 81 (C) Pendleton, Matt, 2000, 01 Pendleton, Tim, 1993 Perkins, Dave, 1989, 90 Perkins, Harry, 1916 Perozzi, David, 1995, 96 Perrone, Mel, 1941, 42, 46 Perry, Bob, 1966 Perry, George, 1936, 37 Person, Robbie, 2008 Peters, David, 1983, 84, 85 Peterson, Dave, 1947, 48, 49 Phillips, Mark, 1988, 89, 90 Phillips, Wallace, 1903 Pianto, Jerry, 1987, 88, 89, 90 Pierce, Jesse, 1999 Piety, Jeff, 1978 Piko, Jim, 1994, 95, 96 (C) Piko, Joe, 2000, 01, 02 Pinkston, Chad, 1999 Pittman, Joe, 2007 Place, Ashley, 1999, 00, 01 Pollizotto, Sam, 1930 Popa, Ryan, 1994, 95 Poremba, Matt, 2000, 01, 02, 03 Poss, Jeff, 2008, 09 Potter, Wally, 1941, 42, 46 (C) Powell, Ames, 1959 Powell, Fred, 1970, 71, 72 Powell, Zane, 1940, 41 Prather, Brad, 1983, 85 Presecan, Nick, 1937 Pritchard, Brian, 1990 Pryor, Dave, 1971 Puchley, Tom, 1982, 83 Puett, James, 1928, 29, 30 Purichia, Joe, 1964, 65 Purkhiser, Bob, 1938, 39, 40 Puskas, Steve, 1956 Q Quale, Dan, 1974 Quiesser, Tim, 1975, 76 Quigg, Ron, 1964 Quiroz, Jordan, 2005, 06, 07, 08 R Raber, Nelson, 1930, 32 Rabold, John, 1937, 39, 40 Rains, Darrell, 1971 Ramsey, Robert, 1954 All-Time Letterwinners Rashevich, Steve, 1991, 92, 93 Ratliff, Vern, 1960, 61 Ray, Cecil, 1931, 32, 33 (C) Ray, Jack, 1887 Read, Scott, 1975, 76, 77, 78 Reagan, Kevin, 2000, 01 Reddle, Scott, 2001, 02, 03 (C) Redmon, Bud, 1887 Redmond, Tom, 1970, 71, 72 (C) Reed, Dave, 1969 Reed, Dick, 1967, 68, 69 Reed, Jason, 1997, 98, 99 Reese, Elgin, 1991, 92, 93, 94 Reichel, Louis, 1922, 23, 24, 25 (C) Reiff, Todd, 1985, 86 Reisler, Phil, 1939 Renie, Tim, 1961, 62 Renner, Jack, 1949 Renners, Randy, 1986, 87, 88 Reno, John, 1940 Renschen, Randy, 1996, 97 Reynolds, Cleon, 1928, 29, 30 Ribordy, Mark, 1984, 85, 86 (C) Richmond, Warren, 1967, 68, 69 Riddle, John, 1951, 52 Ridley, Jordan, 2009 Reigle, Charles, 1967, 68 Ringer, Jim, 1957, 58, 59 (C) Roach, Robert, 1980, 81, 82, 83 Robbins, Patrick, 2003, 04, 05 Roberts, Bob, 1939, 40, 41 (C) Roberts, Dick, 1957, 58, 59 Roberts, Don, 1949 Roberts, Steve, 1986, 87, 88, 89 (C) Roberts, Tim, 1989 Robertson, Troy, 1986 Rodick, Don, 1949, 50 Rodick, Joe, 1941 Rodman, Mark, 1977, 78 Roeder, Jim, 1990, 91, 92 Roehling, Todd, 1989, 90 Roembke, Ron, 1986, 87, 88, 89 Rohrabaugh, Tom, 1954, 55 Romanowski, Paul, 1990, 91 (C) Rooney, Pat, 1987, 88, 89 Rose, Brandon, 1995, 96, 97 Rosenstihl, James, 1947 Rosner, Barney, 1964 Ross, Larry, 1954 Rossell, Keith, 1991, 1992, 93, 94 Rossell, Tim, 1991, 92 Rothhaar, Karl, 1975, 76 Rowley, Mike, 1956 Roy, Curt, 1980, 81, 83 (C) Royce, Francis, 1928, 29 Rudd, Donald, 1938, 39 Rudnicky, John, 1941 Rufli, Lewis, 1929, 30 Runyon, Robert, 1948, 49, 50 Rush, Joel, 2000, 01, 02 Rush, Mike, 1978, 79, 80 Russell, Kevin, 1995 Ruth, Greg, 2005 Rykovich, Bob, 1971, 72 Rykovich, Tom, 1969 S Saam, Tim, 1988, 89 Sadler, Steve, 1964, 65, 66 Salvi, Chris, 2008 Sanders, Brian, 1992, 93, 94 Sanders, Naim, 1995, 96, 97, 98 (C) Safford, Bob, 1951, 52 Sales, Andy, 1983 Sales, Tony, 1981, 82, 83 Sayler, Tom, 1964, 65 (C) Schaffer, Jim, 1975, 76 Schankerman, Maurice, 1949, 50 Scheller, Tom, 1984, 85 Schluge, Lee, 1972, 73, 74 Schluge, Phil, 1971, 72, 73 (C) Schmidtz, Ryan, 2005, 06, 07, 08 Schmitz, Harold, 1971 Schofield, Byron, 1935, 37 Schopf, Robert, 1928, 29 Schuesler, John, 1949, 50 Schuler, Mike, 1991 Schultz, Steve, 1986, 87 Schwanekamp, Brent, 1999, 00, 01 Schwanekamp, Chuck, 1974, 75, 76, 77 Schwecke, Joe, 1977, 78, 79, 80 (C) Schwingendorf, Kyle, 1998 Scifres, Bruce, 1976, 77, 78 Scola, Steve, 1999 Scott, Charles, 1945 Scott, Taylor, 2007 Seal, Mickey, 1960, 61 Sebo, Eric, 1984, 85 Secrist, Ryan, 2007, 08 Seidl, Mike, 2004, 05 Sells, Tom, 1958 Sewards, Bill, 1998 Shafer, Ron, 1957 Shaffer, Andy, 1998, 99 Shaffer, Todd, 1990, 91, 92, 93 Shanteau, Craig, 1977 Sharp, Steve, 1984, 85, 86 Shaw, Aaron, 2006, 07 Shaw, Scot, 1975, 76, 77, 78 Sheehan, Dan, 1955 Sheley, Craig, 2001 Shelton, Derek, 1991 Shepherd, Jim, 1960 Sheridan, Hansel, 1960, 61, 62 Sherwood, Dick, 1947 Shibinski, Mike, 1977, 78, 79, 80 (C) Shirey, Dan, 1987, 88, 89 Shomber, Kevin, 1987, 88, 89 Shook, Larry, 1960, 61, 62 Shultz, Jerry, 1960, 61 Siefert, Mel, 1983, 84 Simpson, Ralph, 1933, 34 Skaggs, Tyler, 2009 Skirchak, John, 1958, 59, 60 Slama, Brett, 1999, 00 Slatter, Anthony, 1998, 99 Sleet, Tom, 1941, 42, 46, 47 Smiley, Ephraim, 1970, 71 (C) Smith, Parker, 2000, 01, 02, 04 (C) Smith, Wayne, 1983, 84, 85 Smock, Ken, 1947, 48 Smothers, Joe, 1967 Snyder, Dan, 1971 Snyder, Leroy, 1945 Sohl, Charles, 1930, 32 (C) Sorrentino, Joe, 1977, 78, 79 Southern, John, 1925 Speron, Roman, 2000, 01, 02 Sporer, Albert, 1937, 38 Spraetz, Ken, 1956, 57, 58 Springer, Bob, 1903 Springer, Dennis, 1989, 90, 91 (C) St. Clair, Steve, 1976, 77 Stahl, Jason, 1992, 93, 94 (C) Stahley, Wayne, 1970, 71, 72 Stalcup, Bill, 1937 Staller, Eldon, 1934, 35, 36 Stallings, Ramon, 1992, 93, 94 Staniewicz, Mike, 2007, 08, 09 Stayer, Tom, 1975, 76, 77 Stearns, Jeff, 1973 Steinmetz, Mark, 1965, 66 Stephenson, Mawrie, 1920 Stermont, Winifred, 1911, 12 Stevenson, Tom, 1894, 95 Stewart, Donald, 1940, 41 Stewart, George, 1965 Stewart, James, 1931, 32, 33 Stewart, Kent, 1957, 58, 59 Stewart, Pete, 1962 Stewart, Robert, 1932, 33, 34 Stewart, Weylin, 1990, 91, 92 Stockslager, Walt, 1957, 58, 59 Stoddard, Eli, 1995, 96, 97 (C) Stone, Robert, 1945 Stout, Waldo, 1934, 35, 36 Stoyko, Steve, 1940, 41, 42 Strahl, James, 1928, 29 Stratman, Rick, 1983 Stratton, Bill, 1945 Straub, Robert, 1949 Streiff, Rick, 1980, 81, 82, 83 Strickland, Richard, 1921, 22, 23 Strole, Gerald, 1922, 23, 24, 25 Stropes, Claude, 1940 Stryzinski, Bob, 1959 Stryzinski, Ron, 1981, 82, 83, 84 Stubbs, Dwayne, 1989, 90 Stump, Jeremy, 2000, 01 Sturgeon, Harland, 1950, 51 Sturm, Don, 1958 Sullivan, Joe, 1930 Sullivan, Logan, 2008, 09 Summerlin, Harold, 1911, 12 Summerville, Spencer, 2006, 07, 08, 09 (C) Sutphin, Dave, 1964, 65 Sutphin, Karl, 1934 Swager, Ralph, 1938, 39, 40 Sweet, Eric, 1979 Sweet, Jeff, 1986, 87 Swift, Cliff, 1934, 35, 36 Swihart, Dave, 1972, 73, 74, 75 (C) Sykes, Lamar, 2003, 04 Sylvester, Bill, 1946, 47, 48, 49 Sylvester, Bill, 1981, 82, 83 Sypult, Chuck, 1982, 83 Sypult, Gene, 1950, 53 T Taber, Matt, 2002, 03, 04 Talarico, Sam, 1988, 89 Tanner, Gordon, 1942 Tapscott, Ralph, 1912 Teague, Frank, 1926 Teague, Jeff, 1986, 87 Teare, Ross, 2008, 09 Templeton, Harold, 1930 Tennant, Jace, 2009 Thaung, Gunner, 1925 Thein, Ernie, 1983 Thomas, Cullen, 1908, 09 (C), 11 (C) Thomas, Larry, 2008, 09 Thomas, Ryan, 2002, 03 Thomas, William, 1933, 34 Thompson, Bill, 1966 Thompson, Brad, 1996, 97, 98 Thompson, Ed, 1894, 95, 98 Thompson, Ed, 1977, 78 Thompson, Leroy, 1953, 54, 55, 56 Thompson, Phillip, 1934, 35 Thompson, Terry, 1979, 80, 81, 82 Thompson, Wes, 1962, 63 Timm, Adam, 1996, 97, 98 Tinder, Ed, 1969, 70 Tiscusan, John, 1939 Toelle, Lowell, 1939, 40, 41 Toner, Chris, 1992, 93, 94 Toner, Dave, 1970 Toon, Herod, 1945 Toran, Derrick, 1986 Torchia, Bill, 1964, 65 Torrence, Steve, 1980, 81, 82, 83 (C) Trabant, Nick, 1998, 99, 00, 01 Tracey, Justin, 2004, 05 Trott, Edward, 1934, 35, 36 Trujillo, Ricky, 2006, 07, 08, 09 Tucker, Albert, 1912 Turley, Ryan, 1994 Turner, B. J., 1996 Tyson, Rick, 2004, 05, 06 (C) U Uhl, Steve, 1992, 93 Ulery, Buck, 2004, 05, 06, 07 Unser, Emil, 1939 Updegraph, Hughes, 1921 Urick, Tom, 1995 V Vandawark, Floyd, 1916 Vandermeer, Mel, 1937, 38, 39 VanDeursen, Matt, 2002, 03 Veith, Grant, 1997, 98, 99, 00 Vermilion, Aaron, 1990, 91, 92 (C) Vermilion, Ryan, 1994, 95, 96 Villani, Mark, 1991, 92, 93 Vlasic, Jerry, 1957, 59 Vonderhaar, Rich, 1971 Vorndran, Phil, 2003, 04, 05 Vosioh, Channing, 1938, 39 Voss, Eric, 1990, 91, 92, 93 W Waggoner, Brandon, 2003 Wagoner, Fred, 1916 Walker, Tim, 1990, 91 Wallace, Brian, 1978, 79, 80 Wallace, Jim, 1967, 68, 69 Wallace, Tom, 1979, 80, 81, 82 Walley, Carter, 2009 Walls, Wayne, 1951, 52 (C) Walsh, John, 1928, 29 Walsman, Bob, 1967, 68 Walsman, Tom, 1970 Walters, Denny, 1965 Walton, Jesse, 1997, 98, 99 (C) Ward, Chuck, 1988, 89, 90 Ward, Kevin, 1996, 97, 98 (C) Warfel, Dan, 1964, 65, 66 (C) Warne, John, 1981, 82, 83 Warrenburg, James, 1947, 49 Watford, Alonzo, 1928, 29 Wathen, Ron, 1955, 56, 57 Watkins, Chris, 2000 Watkins, Zach, 2008, 09 Watson, Marlon, 2000 Webb, Adam, 2003, 04 (C) Webb, Ryan, 2009 Weber, Lou, 1968 Weesner, Ron, 1957 Weger, Jake, 1935, 36, 37 Weger, Ralph, 1932, 33 Weidekamp, Flavian, 1948, 49, 50 Weiss, Mike, 2004 Wells, Charles, 1963, 64 Wells, Rusty, 1981 Welton, Frank, 1936, 37, 38 Wenzler, Morris, 1961, 62, 63 Westerlind, Kyle, 2001, 02, 03, 04 Western, Bryan, 1999 Wetzel, Andy, 1972, 73, 74 Wetzel, Tom, 1977, 78, 79 Wheeler, James, 1937, 38 Wheeler, Sam, 2000 Wheeler, Tom, 1953, 54 Whisner, Phil, 1969, 70 White, Alex, 2006, 08 White, Robert, 1956, 57, 58 White, Ron, 1990, 91, 92, 93 Whitfield, Dave, 1989 Whitt, Carol, 1971 Wiggins, Dionte, 2008 Willett, Brandon, 1999, 00, 01 Williams, Andy, 1941 Williams, Brandon, 2003 Williams, Jason, 1996, 98, 99, 00 Williams, John, 2001, 02 Williams, Norm, 1941, 42, 46 Williams, Orville, 1946, 47 Williams, P. K., 1987, 88, 89 Williams, Rasheed, 1993 Wilms, Larry, 1968, 69, 70 Wilson, Jesse, 1996, 97, 98, 99 Wilson, Michael, 2009 Wilson, Norm, 1952, 53, 54, 55 Windsor, Nick, 2003 Winings, Nick, 1995, 96, 97 Winnings, Ben, 1947 Winters, Larry, 1992, 93, 94 Wise, Glen, 1912 Wise, Verl, 1912 Witman, Bob, 1973, 74 Witmer, Tim, 1989, 90, 91, 92 Wolfe, Richard, 1929, 30 Wood, A. C., 1917 Woodring, Homer, 1924, 25 Woodring, Bob, 1991 Woods, David, 1940 Woods, Gerald, 1921, 22, 24 (C) Wood, Shawn, 1995, 96, 97, 98 Worth, Willard, 1928, 29 Woznick, John, 1951 Wray, Evan, 2008 Wright, Tom, 1947, 48 Wrona, Al, 1975, 76 Wuest, Joe, 1937 Wukovits, Vic, 1966, 67 (C) Wulle, James, 1934, 35 Wymer, Jack, 1959 X Xander, Peter, 2007, 08, 09 Y Yeoman, Todd, 1986, 87, 88 Young, Andrew, 1988, 89, 90, 91 Z Zavela, Dan, 1939, 40, 41 Zavela, George, 1940, 41 Zentz, Tom, 1966 Zimmer, Zach, 2004, 05, 06, 07 Zimmerman, George, 1931, 34 Zimmerman, Scott, 1997, 98, 99 Zimpleman, Ryan, 1998, 99, 00 (C)--Captain. 39 Butler University The five faculty and 113 students present when Butler University opened in 1855 laid a solid foundation for 155 years of creative change and progress. Today’s more than 4,000 students continue to look ahead while treasuring the traditions unique to Butler. The young school, originally named North Western Christian University, was unusually innovative. It was the first in Indiana, and only the third in the nation, to admit women on an equal basis with men. With the appointment in 1858 of Catherine Merrill as Demia Butler Professor of English, the institution became the second in the country to appoint a woman faculty member, the first to establish an endowed chair specifically for a female professor and the first to establish a professorship in English literature. The school was also the first in Indiana to allow its students, with parental consent, to choose subjects under a new “elective” system. As Indianapolis grew, the city’s commercial district began to penetrate the heavily-wooded campus at what is now the corner of Thirteenth Street and College Avenue. In 1873, the board of directors decided to sell the downtown campus and accept a gift of 25 acres in Irvington, then a suburb east of Indianapolis. In 1877, North Western Christian University became Butler University, taking the name of its founder, Indianapolis attorney Ovid Butler. Butler moved again 50 years later, as the “Circle City” continued to grow. In 1928, classes were held for the first time in Jordan Hall, an imposing new Gothic structure erected on the beautiful Fairview Park Site, a wooded tract north of the city on the White River and the Inland Waterway Canal. Today’s students come from nearly every state in the nation and from many foreign countries to enroll in degree programs in the College of Liberal Arts and Sciences, or in one of four professional colleges -- Business Administration, Education, Fine Arts or Pharmacy and Health Sciences. Butler is one of only 21 private schools in the country offering a pharmacy program. True to the vision of its founders, the University continues to offer an array of professional and pre-professional programs within the context of a strong commitment to the traditional arts and sciences and to the values of liberal education. Butler continues to welcome highly motivated, intellectually curious men and women, and to prepare them for lives of professional and community service and creative, ethical action. Butler is one of the top 20 U. S. colleges for producing business executives, is in the top 10% for preparing future Ph.D.s and is located in the #2 city for college graduates starting a career. Administrative Support 40 Barry Collier Director of Athletics Tom Crowley Associate A.D. for Internal Operations Beth Goetz Associate A. D. for Administration/S.W.A. Jim McGrath Associate A.D. for Communications Mike Freeman Associate A.D. for External Operations Joe Gentry Director of Corporate Sponsorship Sonya Hopkins Assistant A.D./Student Development Carl Heck Assistant A.D./ Facilities and Events Stephanie Martin Athletic Business Manager Ryan Galloy Director of Sports Medicine Lindsay Martin Manager, Marketing and Promotions Matt Harris Manager of Fan Development Josh Rattray Assistant Sports Information Director Kyle Smith Assistant Director of the Bulldog Club 2010 Bulldogs, front row, left to right: Mark Giacomantonio, Tadd Dombart, Josh Dorfman, Eddie McHale, Tyler Skaggs, Donnie Gilmore, Mike Staniewicz, Rob Hobson, Matt Kobli, Ryan Secrist, Rob Cosler, Logan Sullivan, Andrew Pratt and Scott Gray. Second row: Trae Heeter, Caleb Conway, Larry Thomas, David Thomas, Christian Eble, David Burke, Brandon Grubbe, Andy Dauch, Greg Egan, Matt Foor, Chris Burns, Steven Depositar, Jordan Ridley, Arthur Monaco, Don Stewart, Daniel Wilson, T. J. Lukasik, Ryan Hitchcock and Jordan Koopman. Third Row: Chris Tinkey (Athletic Trainer), Ted Pajakowski (Student Manager), Stephen Blowers (Student Manager), Grant Lewis (Student Assistant Coach), Jim Peal (Strength and Conditioning Coordinator), Assistant Coach Nick Tabacca, Assistant Coach Matt Walker, Assistant Coach Joe Cheshire, Assistant Coach Danny Sear, Head Coach Jeff Voris, Assistant Coach Nick Anderson, Ryan Galloy (Head Athletic Trainer), Assistant Coach Chris Davis, Assistant Coach Rob Noel, Assistant Coach Tim Cooper, Assistant Coach David Kenney, Thorn Murphy (Student Assistant), Lester Burris (Video Coordinator), John Harding (Equipment Manager) and Missy Schultz (Athletic Trainer). Fourth Row: Andrew Cottrell, Derek O’Connor, JT Mesch, Joseph Purze, Jack McKenna, Dan Haber, Sean Grady, Bill Bork, Bob Olszewski, Logan Perry, Bobby McDonald, Brendan Shannon, Will Schierholz, William Lamar, Kyle Jachim, and Paul Yanow. Fifth Row: Artis Hailey, Matt Hittinger, Jeff Larsen, Zach Watkins, Robert Koteff, Nick Nykaza, Jay Brummel, Jeremy Stephens, Alex Perritt, Joseph Ciancio, David Lang, Michael Wilson, Nick Caldicott, Jace Tennant, John Cannova, Jayme Szafranski and Jimmy Schwabe. Sixth row: Matt Benson, Kevin Cook, Bryce Berry, Tom Judge, Cal Blair, Taylor Clarkson, Carter Walley, Scott Harvey, Andrew Huck, Mike Rose, Paul Sciortino, Nick Schirmann, Ross Teare, Phillip Powell, Charles Perrecone and Stuart Harvey. Top row: Jeff Poss, Ryan Myers, Doug Petty, Matt Storey, Jeff Urch, Grant Hunter, Jay Howard, Pete Mattingly, Charlie Schmelzer, Wade Markley, Mike Wendahl, Brett Thomaston, Nick Atkinson, Ryan Webb, Taylor Harris, Greg Ambrose and Dylan Johnson. Directions To Butler Bulldogs on the Air FROM NORTHWEST: I-65 south to I-465 east, exit I465 at U.S. 31 (Meridian Street), follow Meridian Street south to 49th Street, turn right and follow 49th Street to Butler Bowl/Hinkle Fieldhouse (5 blocks) FROM NORTH: U.S. 31 (Meridian Street) south to 49th Street, turn right and follow to Butler Bowl/Hinkle Fieldhouse (5 blocks) FROM NORTHEAST: I-69 south (becomes Hwy 37 in Marion County) to 46th Street, turn right and follow 46th Street to Boulevard Place, turn right and follow to Butler Bowl/Hinkle Fieldhouse (3 blocks) FROM EAST: I-70 west to I-65 north, exit at Meridian Street (U.S. 31) and follow Meridian Street north to 49th Street, turn left at 49th Street and follow to Butler Bowl Butler Bowl/Hinkle Fieldhouse (5 blocks) FROM SOUTH: I-65 north, exit at Meridian Street (U.S. 31) and follow Meridian Street north to 49th Street, turn left at 49th Street and follow to Butler Bowl/Hinkle Fieldhouse (5 blocks) FROM WEST/SOUTHWEST: I-70 east to I-465 north, exit I-465 at 38th Street, follow 38th Street east to Clarendon Road (first street left after Michigan Road, approximately 7 miles from I-465), turn left at Clarendon Road and follow to Butler campus (7 blocks, Butler Bowl/Hinkle Fieldhouse is located on the northeast corner of the Butler campus) For the sixth straight season, Butler football games will be broadcast on XL 950 (AM). Play-by-play duties for the Bulldogs again will be handled by Brian Giffin, and he’ll again be joined in the booth by former Butler head football coach Ken LaRose. A former collegiate and semi-pro football player, Giffin has covered the Indianapolis Colts, Atlanta Falcons, Tennessee Titans and Tampa Bay Buccaneers for various radio stations and networks across the United States, and he has broadcast high school football in Indiana and Florida. He currently serves as Executive Producer/Engineer for the Atlanta Braves Radio Network. LaRose boasts more than two decades of Butler football experience as a player, assistant coach and head coach. He Brian Giffin Ken LaRose guided the Bulldogs for ten seasons, 1992-2001, and he was at the helm when Butler made the move from NCAA Division II football to the NCAA I-AA. A former Butler football Most Valuable Player, LaRose had the distinction as a coach of leading Butler to league championships in both Division II and Division I-AA. Until the end of the major league baseball season, Butler’s radio play-by-play duties will be handled by Brian Scott, former Butler baseball play-by-play announcer and Sports Director for Radio Brownsburg. Media Information WORKING CREDENTIALS: Requests for working press credentials should be directed to the Butler University Sports Information Office in advance of the game. PHOTOGRAPHERS: Sideline passes are issued upon written request to certified news photographers. Requests should be made at least one week prior to the game. No one is allowed on the field without a sideline pass. RADIO BROADCASTS: Space for up to four broadcasters is provided in the visiting radio booth at the Butler Bowl. One courtesy phone line is available for the designated “Official” radio station of the visiting team. Please direct any additional or special broadcast requests to the Butler Sports Information Office. MID-WEEK INTERVIEWS: Every effort will be made to accommodate media requests for interviews during the week. Players will be made available as class schedules permit. Coaches are usually available prior to noon, or immediately following practice. All interview requests should be directed to Butler’s Sports Information Office. For additional information, please contact: Jim McGrath Sports Information Director, Butler University Office - (317) 940-9414 Fax - (317) 940-9808 jmcgrath@butler.edu 41
http://issuu.com/buathletics/docs/10bufb_guide
CC-MAIN-2015-48
refinedweb
36,008
78.79
I originally posted this article on BlogMarq, but felt it might be useful for more people to see. I used this solution in my #Top10 Windows Phone App #Notebox. Internationalization doesn’t have to be that difficult. I have recently internationalized #Notebox using the following method. First, I added a helper class that will act as my store for the current culture strings/phrase for use in my View bindings: public class InternationalizationHelper { private static readonly AppResources LocalizedResource = new AppResources(); public AppResources Resources { get { return LocalizedResource; } } } Next, I made this class available to my Views by adding this line to my App.xaml: <Code:InternationalizationHelper x: Now for the actual strings and phrases. I added a Resources file called “AppResources.resx” to my project and started adding the Name/Value strings. E.g., Name="LoginPageTile", Value="Login Page". This default resources file will be used by the application Neutral Language, which can be found on the Project Properties, Assembly Info panel. To add further support for other languages, I needed to do twp things: Name="LoginPageTile", Value="Login Page" SupportCulture <SupportedCultures>it-IT;el-GR;es-ES;</SupportedCultures> For this to work, all of the resource files must use the same key name(s). If a resource file for a particular culture does not contain the key name then the default AppResources.resx name entry will be used. The final step was to refer to these names within the Resources file. This was achieved using a binding like this: Text="{Binding Path=Resources.LoginPageTitle, Source={StaticResource localizedResx}}" This article, along with any associated source code and files, is licensed under The Code Project Open License (CPOL) Math Primers for Programmers
http://www.codeproject.com/Tips/381666/Windows-Phone-Internationalization
CC-MAIN-2013-20
refinedweb
280
55.24
Implementing SOLID and the onion architecture in Node.js with TypeScript and InversifyJS Remo H. Jansen on April 10, 2018 In this article, we are going to describe an architecture known as the onion architecture. The onion architecture is a software application archite... Read full post So just an FYI - the project that originated the definition, I was the tech lead on, and we ripped out this style of architecture after about 6 months. For us it didn't scale with complexity. I talked about what we moved to at NDC vimeo.com/131633177 The long and short of it - layers don't encapsulate, they abstract. When we moved to CQRS-style architectures, just the object models that is, we were able to ditch all those service/repository layers in favor of queries and commands. The result was truly SOLID, as opposed to our onion architecture, which I finally saw as a gross misunderstanding of SOLID and OO principles. I just watched your talk and maybe I'm getting it wrong but it seems like you are not happy about the n-tier architecture, not the onion architecture. They are actually quite different (I have also experienced the n-tier pain). n-tier onion More info at blog.ploeh.dk/2013/12/03/layers-on... One of the things you mention is that you are able to change the used ORM tool. In the onion architecture, this is perfectly possible. In fact, we migrated from sequelize to TypeORM and it was a small job. When a system becomes more complex, the onion architecture can be combined with the CQRS and the Unit of work patterns and when it becomes even more complex it can be split into "a bag of onions". No, this was an onion architecture, your second picture, not n-tier. Like literally the dude that invented the term, this was the project and we moved away after a few months. So we start with CQRS and only add any kind of layering like onion as necessary. It's just been 8 years going CQRS-first, and have yet to really need any layers. No abstractions, nothing like a repository or service layer, nothing like that. Just unnecessary. "The long and short of it - layers don't encapsulate, they abstract." Effing nail on the head there mate. We've been slowly massaging our front end away from onion because of exactly that over-abstraction to the point of sprint velocity crushing complexity. It also lead to a lot of premature generalization. I saw your video and seems its like coming back to square one after all the DDD, that that patterns, just simple separation of concern . In one of my earlier project I was doing dividing the project based on the controller, model and service structure, then in new version I thought the feature separation will help, however eventually I felt that feature separation was making me more concerned as where to keep some files or function. Like some functions need to be dependent on two or more features, where to create that folder structure, then many a times comes a lot of cross dependability, how those things can be taken care? In that case I feel that a particular functionality instead of feature works better. Congratulations, you have now turned JavaScript into a shittier version of Java. I have had this discussion before with friends that started on dynamic languages and are now enamored with types and the SOLID principles. Dependency injection and nonsense like 300 interfaces are the things that drove half of the Java community to move to Ruby and Python 10 uears ago. I personally love JavaScript for it's practicality and adaptness. Tieing it up with types is bad enough, and if you add DI and interfaces all around the place, it loses everything that makes JS more interesting as a language than Java or C#. Coming up next, Fizz Buzz Enterprise Edition in Typescript, because serious node.js code is written by serious businessmen to support serious businesses. Honestly though, I recommend you read Eloquent Ruby to learn new idioms that make sense on dynamic languages. I respect your point of view but my personal experience is radically different. When I have something like the following: There are a few things that I wouldn't get without types and without dependency injection: I can do TDD / design by contract with confidence. If I'm implementing a type, the first thing that I can do is click on "Implement interface" on my IDE and it will create a default implementation. If my implementation is a violation of the interface I will get a compilation error. While I implement this controller, another engineer in my team can be creating an implementation of AircraftRepository. We can work in parallel and we know that there will be no integration issues because the interface is a contract between us that cannot be violated by either side. I can use declarative routing @controller("/api/v1/aircraft")without the need to implement a factory for my controller by hand. I have a lot of metadata that can be used for things like generating UML diagrams or Open API definitions. If I need to refactor something I can do it with confidence, a change in the contract will break all the related parts. I know exactly what is left for my refactoring to be completed. I can also use InversifyJS to implement interception (This is not documented in the example but it is supported by inversifyJS). This allows implementing an aspect (e.g., logging) while keeping my entire code base free of logging concerns. These advantages make a difference for me and I enjoy them. This is why I advocate this approach and I'm sure I'm not alone because there are a lot of emerging Node.js libs and frameworks in this space and some of them are becoming very popular (e.g, nestjs.com/). I can agree that if you are used to that way of working and actually enjoy it (personally I'm a J2EE refugee and enjoy the fact that I've left that world of pain), sure, have a blast. But the thing is that almost every meaningful feature in this approach already has a simpler, more streamlined and more idiomatic alternative in Javascript. For example, DI is useful for testing. But there are testing-oriented DI implementations for JavaScript that do it much nicer than the square-peg-round-hole approaches like InversifyJS. I'm talking SinonJS, Testdouble.js and Steal.js. There are numerous AOP solutions for JavaScript like Aspect.js or Meld that are both simpler and more idiomatic than DI inception based ones and things like Express' Debug clearly demonstrate that where AOP is needed it can be added without needless OOP overengineering and it's actually the anyithing-goes, dynamic nature of the language/runtime that massively helps in this regard (higher order functions alone make a world of difference). In actuality an API is a contract. OO interfaces are really just one way to go about them and not necessarily a better one. But as IDE support is a given using interfaces, I can imagine circumstances where that might be preferable to documenting an API or adhering to agreements, but IRL I've never run across such a situation, yet I've dealt with incredibly slow progress and numerous pain-points in OO-heavy codebases using these patterns (despite working with stellar engineers). I'm not even interested in going into a debate on typing. The thing is that there are no conclusive proofs that types reduce bugs at all (but abandoning OOP for functional programming apparently does), and one can sprinkle Typescript or native type annotations where one finds them needed without full-on commitment (where lack of strict types has actually helped dynamic languages and JSON to flourish in this world of distributed software and fast-changing APIs). The "without the need to implement a factory" and "generating UML diagrams" bits in your comment do say a lot too. Just my 2c. History lesson...Microsoft Windows is built as a bundle of exposed API's -- unlike linux with text config files -- and they caused immense pain in the 90's by breaking their bogus contracts. So I beg to differ, API's are not real contracts. In fact I'm sure the heinous behavior of Microsoft's APIs were one of the reasons Java made interfaces a big part of its paradigm. Take it easy. Here, in the article, there´s some work and time invested trying to explain another point of view about node and how to build software in this platform. So, respect for other opinions. I respect what you`re saying about JS and how people work with it taking advance of the flexibility that provides...but this requires a very advanced skills and discipline to maintain the state of the soft clean, stable and readable for others. You must be really good on JS (sure you are), but for others like me who are not that proficient on JS we appreciate solutions like this or Nest that allow us to build (yes, the same house, with the same structure, features etc many times...) projects safely, with rail guidance etc because sometimes you can not wait to be a rock and roll programming star to build and finish an app. With JS you can build custom fancy houses, great... but sometimes you have to build a block hundred identical houses with many workers and in this scenario, Remo solution, fits very well in terms of productivity. I think that it was not Remo´s intention to offend "node's pures engineers" ;) So you will probably hate DENO too Hi Alex, I sort of agree with your view. However could you suggest some alternatives patterns to better structure node apps? First of all, great post! But, i can't understand why you shouldn't use java instead of convert javascript into something similar to java. I mean, i can understand why use ts in frontend since you don't have the java possibility.. but, why should i use it in backend? And i'm just asking this because i want to understand and share ideas :) Well, I can enjoy some of the things I like about Node.js / JavaScript / TypeScript: But I also get some of the benefits of Java / C#. I had many conversations with many engineers and some of them are completely against things like dependency injection in JavaScript. They argue that it is not needed because in JavaScript you can monkey patch everything if you need to mock something during a unit test and the same applies to languages like Ruby. For example, you can read this post. On the other hand, you have people like me or Rob Wormald: This is cool for me, after a long time thinking about this topic (I'm quite passionate about this topic), I have reached the conclusion that it is not right or wrong. If monkey patching works well for you and your team that is awesome. I stand on the other side: monkey patching and lack of static types don't work well for me. I appreciate questions like yours because as I said I love thinking about this topic :) When I wrote this article I wasn't saying "This is the way you should build Node.js apps", what I was trying to say was "This is the way I've been building very large monolithic apps, with large teams and it has been working very well so far". Yes, I totally understand what you are saying. DI it's pretty cool and it's something that I suffered while I was a working with Ruby. I mean, you can redefine methods or classes at execution time when you need it. But it could be pretty complex at the beginning. Like you say, there's not good or right for this kinds of topics. Thanks for your answer! Really nice article and it's about time to introduce methods like this to the JavaScript developers. Actually I'm not surprised, that some or a lot of JavaScript developers are complaining about it. My experience is that it is very hard for them to understand what design are for and what benefit they have. Especially whe it comes to bigger or more complex applications. It was, and sometimes still is, the same in other languages. For example I know PHP developers who likes global functions, unclean architecture or everything with static calls, or don't use things like namespaces right. Hi Remo! Great post! Thanks for sharing!! Is the source code available somewhere? i.e github, etc? Thanks! Not the one described in this post because is private but I have a small demo is much more simplified and rudimentary but it can help you to get an idea github.com/stelltec/public-tech-de... Wow...Really good work with the blog, finally found an example that totally makes sense. Thank you so much, been searching ages for a decent JS example that ties in with DDD and SOLID. Would love to see the source code for the actual blog. Saw the small demo example and it is very neat. I don't understand why others cant see the benefits of DDD and SOLID. Is there a way to see the private example? Awesome! Thanks! what is the project template did you use in your sample source code? thanks. The code samples are based on some proprietary code that we created at my current job. I cannot share the original source code as I described already. Sorry. thank you for a great article! few points: don't you think it smells that you need to have your IoC framework decorators inside the lowest level layers such as the domain? Uncle Bob in Clean Architecture specifically suggests to avoid that.. the IoC is also just an implementation detail. I couldn't find where you use the "Application Layer". isn't it missing? for example "create new Aircraft" should not be considered a use-case belongs to the Application Layer? As far as I understand, domain services should be introduced when you have a domain business rules. but in your case, you put there operations which don't hold any business rules ("get all aircrafts" for example). I think it indicates that it does not belong there. I even picked at the demo you shared (github.com/stelltec/public-tech-de...) and there is not even a folder for this layer. on this article: blog.codeminer42.com/nodejs-and-go... (you can skip to this code at: github.com/talyssonoc/node-api-boi...) The guy put those operations in the application layer for example. do you think he's wrong? Thank you so much for your comment and for your questions and links! It's really helpful, especially this one (github.com/talyssonoc/node-api-boi...) I think you forgot to add the Serializable interface to the Rectangle class in the Interface Segregation Principle section so that it can use the serialize() method. Thanks a lot for the heads up! Yup yup! Great article by the way! 👍 Extremely nice post! Well explained SOLID and onion architecture with good examples, thanks! Thanks! :) As someone who loves this kind of architecture, It was a great read! I need to look more at Typescript, InversifyJS, and TypeORM. It seems a viable alternative to PHP for OOP projects. (Not that I want to replace PHP in my stack, but its always great to have another option, mostly for some projects that could benefit from the NodeJS ecosystem). Late to the party but really excellent article! I've worked hard in recent times on implementing onion architecture (and later hexagonal architecture) in a strictly FP front end and I can't imagine going back to a code base organised exclusively by domain. For reference and self-plugging I created jpex to achieve dependency injection in a similar vein to inversify but with FP at the forefront. Any one know what is @dal? DAL means "data access layer" and is an alias for a path. Check this out to learn how to create aliases for paths in TS stackoverflow.com/questions/432817... Hi @RemoHJansen, I'm switching from Golang to Typescript, i dont know to discriminate standard library in Typescript. Please let me know how to discriminate standard library. Thanks you! What do you mean by "discriminate standard library"? How do you know what is standard library in Nodejs and Typescript? In Golang i prefer to use standard library(golang.org/pkg/). If in this list there isn't library which i need, i must pick thirt party library. In Node.js and TypeScript everything that you install via npm is not part of the standard library. The standard library is knowns as "Node core" and you can check the available modules (they are listed on the left menu) at nodejs.org/dist/latest-v10.x/docs/... @remo H. Jansen I can't find out when it was implemented and i can't understand where '@dal' point to. Please help me Why do you need the EntityDataMapper? why not use the domain entity directly? I am curious about how did this work with typeorm and how did your DalEntities synced with the database ? We use the data mapper because we don't want to "pollute" our entire application with data-access concerns. The TypeORM entities are data-access concerns because they are a 1-to-1 map to the database tables and contain database-related metadata added via decorators. We use the domain entities instead in the entire system but the data access layer (DAL) cannot use domain entities to persist them (via TypeORM). So the data mapper job is to translate in and out of the DAL from domain-entity-to-orm-entity (in) and from orm-entity-to-domain-entity (out). How do you work with transactions using TypeORM with this architecture? TypeORM seems to want to break Separation of Concerns by using Transaction Decorators. Thanks. Hi Remo H.Jansen I don't know what TYPE.TypeOrmRepositoryOfAircraftEntity is, Can you let me know about it? Thank you! Amazing stuff. Really liked the suggested flow and how it helps in separating concerns. Is there some book to learn SOLID principles?
https://dev.to/remojansen/implementing-the-onion-architecture-in-nodejs-with-typescript-and-inversifyjs-10ad/comments
CC-MAIN-2021-25
refinedweb
3,081
63.29
Cassandra vs. MariaDB 12 min read· Keeping the ‘Hat Shop’ in Business Welcome to the world’s most popular hat shop. Our database is old and creaky and failing to keep up with today’s demand for our brand new tin foil hats. We need a new, more scalable, solution capable of handling hundreds of order updates per second. We’ll be updating our hat orders so our customers can track their orders with values such as cost, colour and quantity of their desired head gear (our customers are very discerning). Like comparing Fedora’s and Fez’s… It’s a foregone conclusion that if horizontal scalability is necessary in a system Cassandra should outperform a traditional relational database. So instead we’ll be looking at how Cassandra performs against MariaDB in a single node setup and how easy it is to use from a development perspective, in particular what the performance of Cassandra is like before we start to scale out. The Hats To test our databases, we need some data. We built a system that can generate events as serialised JSON which can be actioned against our databases. We chose standard CRUD operations for our event types, i.e. creating new orders, reading orders, updating the order status and attributes (colour, size, quantity) and deleting orders. Our Hat Shop relies on Java’s Random library to create a range of events and orders. To ensure that our experiments will be repeatable (that we can run the same data against each database), we seed our random generator so that for each test, it will generate the same set of events and orders. Generated events look something like this: { "type":"CREATE", "data":{ "id":"4612d1c0-212d-40a0-813d-a17543296896", "lineItems":[ { "id":"561cea89-1fd8-4e94-800b-626ca27149f9", "product":{ "id":"414dfe60-3d7f-445c-8b82-7620f6cea367", "productType":"HAT", "name":"Beanie", "weight":594.81, "price":6.29, "colour":"Burnt Sienna", "size":"XXL" }, "quantity":9, "linePrice":56.61 }, { "id":"cde58d2b-0f60-45d4-b46a-94526f001ec4", "product":{ "id":"213a29b3-e411-4684-8925-4aca4a1d1e22", "productType":"HAT", "name":"Flat", "weight":168.9, "price":64.09, "colour":"Puce", "size":"S" }, "quantity":8, "linePrice":512.72 }], "client":{ "id":"4612c308-9158-4bbe-bc41-52b25647a2e6", "name":"Katherine Sims", "address":"240, frantic Drive, Worcester", "email":"KatherineSims@fakemail.com" }, "status":"ORDERED", "subTotal":2779.22, "date":1481537772610 } } …or Cloche’s and Pork Pies Once we had some data coming in we needed to write the code to handle these events. We made a few design decisions along the way in an effort to make a fairer comparison. We considered, for code-readability and to make database setup easier, using Spring with Hibernate to handle accessing MariaDB data but we opted against this. Instead, for more transparency of how database queries are handled and to ensure equivalency in our testing, we created the database queries ourselves. We also decided against using the messaging queue software RabbitMQ to send generated events from the hat shop to the databases. Whilst it was useful in allowing us to separate out the data generator code from the databases, fully decoupling the different components of our application, it limited the speed with which we could throw data at the databases. RabbitMQ also had the advantage that it allowed us to play the same events against both databases, however our hat shops seedable random number generator meant that we had this capability already and so we made the decision to leave it out. The Datastax driver for Cassandra offers asynchronous execution of queries by calling the executeAsync() method. We used Java to run our experiments, so for MariaDB we had to use JDBC for our queries. JDBC doesn’t offer asynchronous execution, so we execute our queries from isolated threads and use a connection pool provided by HikariCP to give use something like the Cassandra driver’s in-built functionality with JDBC. Lastly we wanted both our databases to be run in an isolated environment which could be easily recreated for testing. To this end we made the decision to use Docker images to run both Cassandra and MariaDB enabling us to run up both database instances quickly and easily. The system however, is written in such a way as to easily allow connections to be made to database instances outside of Docker if necessary. Logging our results We wanted to persist the results in a human readable form that was quick to append to and which we could sit a data visualization tool on top of. To this end we logged the following information for each test to .csv files and used Microsoft Excel and PowerBI to produce graphs of our data. - Test ID - Database Type - Event Type - Time Taken - Success (If the query succeeded) - Error Message (If any) - Time Stamp We ran tests using identical sets of events and measuring the total times with and without logging, against both MariaDB and Cassandra instances to see if there was a measurable delay due to logging. Indeed, there was a small effect in processing the log, adding between 1 and 5 seconds for 10,000 events but only for our overall time. We are timing each database transaction individually and these timings are not affected by the logger. For tests measuring overall time, logging is disabled. Thinking cap - Database differences & tweaking performance Cassandra is a noSQL database, designed to give high performance when distributed at scale. While the database uses terms such as ‘row’, ‘column’ and ‘table’ the underlying data structure is different from a relational database and is more like a Map of Maps. Data in Cassandra is stored at the top level in a keyspace, which is a namespace used to define replication across nodes. Within a keyspace are Column families which are analogous to SQL tables. These map row keys to sorted Maps of column keys and values. Cassandra also ships with its own query language CQL (Cassandra Query Language) which is reminiscent of SQL. The similarities in query language and names like table and row can make it tempting to treat Cassandra as an SQL database when it comes to table design and data query however this can lead to an inefficient approach if you’re not careful, as we found out. Join operations are not permitted in Cassandra however Primary Keys can be composed of more than one column’s values. The first element in the Primary Key is used as a Partition Key for the table and the others as Clustering Keys. The Partition Key is used to hash the data across the nodes in Cassandra and as such can have a large impact on the efficiency of queries. To maximise efficiency the goal is to distribute data evenly across nodes and to access as few nodes as possible with a single query. This involves structuring the tables in your database in a way that facilitates this. In Cassandra rather than modifying the data in a table on receiving an ‘update’ query, a new row is created instead. Data in Cassandra is time-stamped and when an update event happens data is written as a new value and the old value is marked with a tombstone and deleted at a later date during the compaction process. This process also helps to increase performance on ‘delete’ events in the same way as Cassandra does not delete the data immediately but simply marks it with a tombstone as with an update. Tweaking To ensure that we treated each database as fairly as possible we amalgamated the pre-processing stage of our code such that both Cassandra and MariaDB share as much of the same setup process as possible. Timing is started just before execute is called on each database and stopped once a result is returned. The program remains running until all threads in the thread pool have finished to ensure that we can calculate a reasonably accurate end to end time for the tests. We also made the decision to implement batch statements and asynchronous execution functionality on both databases in order to optimise the queries made on the databases. Cassandra Tweaks During the testing process we came across a good example of how the table layout in Cassandra can affect the performance of the database. Our initial tests with ‘update’ events consistently showed MariaDB as being the faster database when processing multiple updates. Cassandra registered times of on average between 10 and 16 milliseconds for transactions to complete. Originally the database tables were similar to the MariaDB tables. An ‘orders’ table and a ‘lineItems’ table. The lineItems table was structured with the following fields: line_item_id (PK), order_id, product_id, quantity, price This structure means that the data in the line Items table is distributed across the nodes* in the database using a hash of line_item_id. This is fine for queries on a single line item which require access to just one node but slows performance for querying multiple line items which requires access to multiple nodes. For operations which require querying multiple line items MariaDB makes use of the foreign key order_id to query all rows with the same order_id. In Cassandra however each line item resides potentially on a separate node and so the query becomes inefficient and functionality to perform such queries is restricted by Cassandra to avoid this. To alleviate this problem the line items table was restructured to better suit Cassandra’s distributed nature and renamed line_items_by_order. order_id(Partition Key), line_item_id(Clustering Key), product_id, quantity, price The table now has a composite Primary Key made up of order_id and line_item_id. As discussed the first element (order_id) serves as a Partition Key. This means that now all line items of a given order are stored under the same node in the database, now when a query is made on multiple line items belonging to an order, Cassandra only has to access one node to retrieve all relevant data. Fig 1 shows the times for different queries with the old table style (line_items) and the new (line_items_by_order). As can be seen access times are noticeably improved in most instances. Fig 1 * Or virtual nodes (vNodes) in our case as we are only using a single node instance Configuring a Connection Pool for MariaDB Fig 2 shows the influence of the maximum connections setting on performance in terms of the overall execution time. To produce this chart we took the average of the total execution times of 5 runs of 50,000 ‘create’ operations against MariaDB at each max-connection parameter we tried from 1 to 30. Fig 2 At a maximum connection pool size of 1, operations are effectively running synchronously, taking nearly 2 minutes to complete the operations. At a maximum of 2, the total execution time is reduced to just over 80 seconds. At 4 we stop seeing the benefit of increasing the size of the connection pool further, achieving an execution time of just over 1 minute. The connection pool holds a set of threads, each with an independent connection to the database. The machines upon which we’re running our tests have 4 processor cores meaning that at most, we can only be executing 4 threads at once. Fig 2 shows we could run more, keeping extra threads in memory and switching to them as the cores become free, at no cost to overall time but it would reduce performance in terms of each individual queries response time as the database will respond most quickly when it has the fewest number of clients making requests to it. Changing hats - how easy is it to update an order? Starting with the basics, we wanted to know how easy it was to perform the CRUD operations described above. MariaDB MariaDB has two tables, orders and line_items set up, respectively, with the following fields: order_id (PK), client_id, timestamp, order_status line_item_id (PK), order_id, product_id, quantity There can be multiple line items to each order. When creating an order, two SQL queries are run, one to create the order and another to create the line items associated with it. To create a new order: INSERT INTO order VALUES('1234-5678-9101-1121', '3141-5161-7181-9202', '1234567890', 'Pending'); To add 3 identical line items (primary key is handled automatically) to the order: INSERT INTO line_item(order_id, product_id, quantity) VALUES('1234-5678-9101-1121', '0212-2232-4252-6272', 1), ('1234-5678-9101-1121', '0212-2232-4252-6272', 1), ('1234-5678-9101-1121', '0212-2232-4252-6272', 1); To update the order status: UPDATE order SET status='Out For Delivery' WHERE id='1234-5678-9101-1121'; No need to say more about the queries than this, it’s clear enough how they work. This is exactly in line with what one would expect from an SQL database and came together easily. There’s very little to worry about regarding optimising this further, it’s a standard many-to-one relationship. Cassandra Setting up Cassandra was also fairly straightforward, we tried both a Docker image running in Ubuntu and the Windows MSI, both allowed a working Cassandra instance to be create quickly and easily with a simple command or double click. Once running we connected to the database using the Datastax Cassandra Java driver allowing us to programmatically create a cluster, session and lastly a keyspace. This again was straightforward with the only issue being to ensure the driver version you are using is compatible with the Cassandra version. Once we were up and running, queries were constructed to setup the database by creating our two tables. Writing queries for the database is almost the same process for Cassandra as for MariaDB, the Cassandra driver supports Prepared statements and Batch statements, and the syntax of the CQL is very similar: To create a new order: INSERT INTO order.orders ( order_id, lineItem_ids, client_id, date_created, status, order_subTotal ) VALUES(?, ?, ?, ?, ?, ?); To add line items to the order: INSERT INTO order.lineItems_by_orderId ( order_id, lineItem_id, product_id, quantity, line_price ) VALUES (?, ?, ?, ?, ?); To update the order: UPDATE order.orders SET date_created=?, status=?, order_subTotal=? WHERE order_id=?; As can be seen, from this perspective there is no extra effort involved in querying Cassandra vs MariaDB. Results n Random Events Running 200,000 random events against a relatively small number of orders (500) shows that for ‘create’, ‘delete’ and ‘read’ events both databases are roughly equivalent, managing to return from a query in under a millisecond on average (Fig 4). Reads are on average slightly faster in MariaDB in this configuration but not by much and conversely updates are faster in Cassandra against MariaDB by about the same difference. Updating a single column value however shows a slight difference between Cassandra and MariaDB of on average 2 and a half milliseconds. This may be due to the fact that Cassandra has to access a row to update a column rather than simply write a new row and so it would seem to be more efficient to simply update an entire row with Cassandra. We also found that running these tests with differing sizes of data pre-loaded into the database made very little difference using the simple configuration we set up. n Update Events We ran multiple ‘update’ events against both databases in sizes of 500, 5000, 10,000 and 100,000. The results of this showed that whilst the average time taken for an update to succeed remained a relative constant for each database, in MariaDB the time taken to perform updates is roughly double the time taken in Cassandra and there is less consistency. This is perhaps a result of the fact that in Cassandra updates are treated the same as writes to increase performance. Fig 3 Average Response Times & Overall Execution Times Fig 4 shows the average response times, by event type, for the same 200,000 events against both of our databases. We made use of the connection pool for MariaDB, setting it to a maximum of 4 connections. As explained under ‘Configuring a Connection Pool for MariaDB’ this should theoretically be the ideal setting to minimise the overall length of execution and to make it a more direct comparison to the Cassandra driver’s executeAsync() function. These figures should not be considered as rigorous benchmarks, we attempted to bring the MariaDB driver into line with Cassandra driver by making it work asynchronously with the connection pool but neither database has been optimised beyond default settings. Our single-node setup lets us play with the performance features that are under the control of the developer but neither database would be deployed to production in the way we have it set up. That said, it is still interesting to note that as well as having comparable response times to our single-node instance of Cassandra, the overall execution time of MariaDB was just below 55.7 Seconds compared to Cassandra’s 4 minutes and 30. Fig 4 Conclusion - Hats all folks Whilst we have gathered interesting results it is worth remembering that we are measuring differences between the two databases in terms of milliseconds. The most dramatic changes we have observed in timings during the course of our investigation have been as a result of configuring how we setup and access the databases rather than differences as a result of volume and velocity of data. Introduction of threading, batch queries, connection pools and remodelling table structures have reduced times by as much as 15 milliseconds on average across both databases. Cassandra handles events as fast, if not faster in some cases, than MariaDB in a simple single node setup without scaling the database out, and using a query language and setup which is similar to, and no more complex than a standard relational setup. This however is dependent on how the database is setup and accessed, knowing the queries that are going to be performed and designing tables around them is of great benefit for performance in Cassandra, as is using the Datastax drivers inbuilt asynchronous execution. If you know what queries you expect to make on your database, may need your database to scale out and will benefit from a flexible schema then Cassandra would be a good choice. It is very close in comparison to MariaDB In terms of performance and handles row ‘update’ operations particularly well. If your data fits well into a relational model however and you don’t need to scale horizontally then MariaDB provides all the familiarity and benefits of a relational database and performs particularly well for ‘read’ operations as you would expect.
http://blog.scottlogic.com/2017/03/01/cassandra-vs-mariadb.html
CC-MAIN-2017-17
refinedweb
3,070
55.37
hi,i'm still a beginner in C++, below is a sample program n its output, i would like to know how do I limit the match recorded to only once? Refering 2 the 2nd line output it found a match at array1 element 1 and array2 element2. But it finds the same match again in the 3rd line output at array2 element2. When a match is found, the scanning shouldn't repeat the matcher array1 and the matched at array2, it should proceed for the new next consequent array's element. Thanx to anyone who would be able to shed some light in it.thanx alot. #include <stdio.h> #define MAX 5 int main(void) { int i, j, array1[5], array2[5]; array1[0] = '1'; array1[1] = '0'; array1[2] = '0'; array1[3] = '7'; array2[0] = '1'; array2[1] = '5'; array2[2] = '0'; array2[3] = '2'; for (i=0;i<MAX-1;i++) { for(j=0;j<MAX-1;j++) { if (array1[i] == array2[j]) { printf("we found a match at array1 element %d and array2 element %d \n", i, j); } } } return 0; } Output screen: we found a match at array1 element 0 and array2 element 0 we found a match at array1 element 1 and array2 element 2 we found a match at array1 element 2 and array2 element 2 thanx again
http://cboard.cprogramming.com/cplusplus-programming/42153-2-array-match.html
CC-MAIN-2014-52
refinedweb
224
55.71
Guest Post by Willis Eschenbach The “Annual Energy Outlook” for 2011 is just out from the US Energy Information Administration. The section called “Levelized Cost of New Generation Resources” looks at what are called the “levelized” costs of electric power from a variety of sources. Their study includes “renewable” sources like solar, although I’ve never found out exactly how they plan to renew the sun once it runs out. The EIA data in Figure 1 shows why solar will not be economically viable any time soon. Figure 1. Levelized costs of the different ways of generating power, from the EIA. Blue bars show the capital costs for the system, while red bars are fuel, operations, and maintenance costs. Estimates are for power plants which would come on line in five years. Operation costs include fuel costs as appropriate. Background: HR diagram of stars in the star cluster M55 . So why is this chart such bad news for solar electricity? It’s bad news because it shows that solar won’t become cheap enough to be competitive in the open market any time in the near future. Here’s why. Now, please don’t get me wrong about solar. I lived off the grid for three years on a houseboat with solar power in Fiji, collecting sunshine and drinking rainwater. I am a solar enthusiast and advocate, there are lots of places where it is the best option. But not on the grid. It’s too expensive. Yes, it’s true that the sunshine fuel is free. And the operations and maintenance is cheap, 2 cents a kilowatt-hour. And as backers are always claiming, it’s the only technology where the capital cost is falling rather than rising, as the price of solar cells drops. … That means that out of the twenty cents of capital costs for solar, only about six cents is panel costs. Let us suppose that at some future date solar panels become, as they say, “cheap as chips”. Suppose instead of six cents per kWh of produced power, they drop all the way down to the ridiculous price of one US penny, one cent per kilowatt-hour. Very unlikely in the next few decades, but let’s take best case. That would save five cents per kWh... And that means that the dream of economically powering the grid with solar in the near future is just that—an unattainable dream. The idea that we are just helping solar get on its feet is not true. The claim that in the future solar electricity will be economical without subsidies is a chimera. w. PS—On a totally separate issue, … 317 thoughts on “The Dark Future of Solar Electricity” The latest Nuclear power stations are designed for a minimum life of 60 years and in the case of the new generation Thorium Reactors being designed in India, 100 years, what would be the levelized cost in cents of nuclear power if the length of the operating life cycle was taken into account? Thanks for the cost comparisons, Willis, Your misgivings about the true costs of wind power are most likely justified. I can think of a few things that should be taken into account and possibly were not, Take standby power generation when the wind doesn’t blow (or blows too hard) and the wind turbines are just idling — when conventional generating capacity has to take up the slack, the cost of connecting wind turbines to the grid. Then there is the question of whether the cost estimates for wind power are based on the theoretical maximum capacity rating of the turbines or on the real generating capacity of around 22 – 24 percent of rated generating capacity. Walter H. Schneider says: December 3, 2011 at 1:30 am. Never thought I would disagree with You Willis, but on this I do.. And of the joy to get independent. A small scale revolution. ;-) Walter H. is correct. Now that alternative power plants have had a few years of running some interesting performance figures have come to light. The best the UK can get is around 17% for wind turbine and have their fair share of maintainence issues, network instability etc.. Here in Australia new subsidised solar panels have been running for more than a year. A 10kW solar unit shows that it can deliver around 40kwh per day averaged over 12months. Denmark boasts that 50% of their generation capacity comes from wind turbines, however they only manage to deliver around 15% to 20% of their capacity. A typical coal fired generator can deliver around 90% of it’s 660mW name plate rating 24/7. You will need around 300 wind turbines to match that name plate rating only, but will never deliver 24/7. But you are using logic and facts. That’s not the way the system works. Governments do what is politically in their interest to stay in governemnt. Even in a steadfastly nuclear France, the coming government will be a Socialist/Green combo. To support the Socialist to get enough votes the Greens have demanded that nuclear will be phased out completely. The Socialists have talked that down to downsizing from 80% to 50% by 2050. Totally absurd capital destruction and will cause a tripling of current electricity costs making France even less commercially competitive (if that is even possible ) To counteract this, the current government in the hopes of being reelected has now commissioned a huge offshore windpark. Go figure, your energy is practically 100% clean but lets waste a fortune we don’t have on pointless exercise. All to get reelected. That’s how it works, not if it’s in any way reasonable… coldlynx says: December 3, 2011 at 2:00 am Excellent. Good to hear from you. I wish this were true. But see my post regarding a huge megawatt-scale project in California. The related article said: Note that that is what PGE is paying, so they will have to sell the power for much more than that. You go on to say: I have no problem with that provided I’m not asked to subsidize it. Generally not true. Here are wholesale prices, $1.36 to $2 per watt. But as I pointed out above, that’s far from the whole cost of the system. Also, there’s a huge oversupply of panels at the moment. The Chinese ramped up big, and then subsidies ran out in lots of countries. So panels are cheap, but not likely to get a lot cheaper in the near future, the market is still correcting. Finally, as I said above, panel costs are a small part of the whole equation. The “many places” are generally places like California, where prices are artificially inflated. I know of nowhere that solar is a “good investment” where there is no subsidy for solar. I don’t care if it’s large or small PV installations, the economics are not that much different. In fact, small installations are generally less efficient than large institutions, which means that prices are higher. I admire and enjoy independence as much as the next man. If a person is getting paid by the grid to produce power, however, you need to watch out for the hidden subsidies. You say it is a “good investment, for consumers”. Solar is a good investment in California if a residential customer is paying more than about $160.00 per month. But the reason the customer is being billed that much for their power is because PGE power is so expensive. And why is PGE power so expensive? Because, as the quote above shows, PGE is paying “50 percent more than the expected market cost of electricity in California from a newly built gas-powered plant” for solar generated power. So they have to sell it at a high price … which is what makes solar a “good investment for the consumer” around here. So solar is a “good investment” for consumers, but only because the consumer is already getting screwed to pay for that investment, with the ratepayer shelling out 50% above market for solar power. Thanks, w. Willis, inverters get cheaper, too. Gotta echo coldlynx in that regard. The trend is towards one inverter per module, and through some as yet unforeseen magic, we’ll have modules that we directly connect to the mains in 20 years; the inverter will be integrated into them. I agree with your comments on the maintenance costs for wind. I do research and consulting work in the power industry, coal and gas combined cycle, and based on my experience the distributed nature of the wind turbines will increase the maintenance costs particularly as they age above the EIA assumptions . I think you didn’t consider efficiency improvements. These lower all cost components. Willis: Great piece. I looked at the utilization rate of Wind in Denmark which has very extensive experience over 30 plus years. The average utilization rateover the last 10 years is approximately 22% – by my calculation, i.e., Generated TWh*POWER(10,12)/(Capacity in MW*POWER(10,6)*365*24) This seems to me to be a better estimate – it includes off shore generation capacity as well. (see) What it does’t factor in is distribution costs that are typically add $0.07-0.15 per kWh. In Southern California consumers typically pay $0.2 per kWh or more so PV power produced on your own roof can be pretty competitive. Willis Eschenbach says: December 3, 2011 at 2:34 am ” I have no problem with that provided I’m not asked to subsidize it.” I recently had a PV system installed which is planned to pay back in about eight years. That is based on the existing feed in tariff of 43.1p which is soon to drop to 21p for new installations. Even if the FIT is reduced to zero in the future, there will still be a hidden subsidy in that the rest of the suppliers customers are paying for the back-up supply for when the sun don’t shine. I may not have put that very clearly, but I hope you get the gist. Good post, Willis. Just one thing: when you complain about solar subsidies, don’t forget that fossil fuels are subsidized as well! The two obvious errors (or call them bad assumptions if you wish): * A 34% capacity factor for wind is just wishful thinking * A 30 year cost-recovery period overestimates the cost of both hydro and nuclear The 30 year period might also effect other energy sources. However the relation between capital cost and the running cost of fuel and maintenance, and their >30 year life time, makes the choice of a 30 year period effect hydro and nuclear the most. A couple of points. I live in perhaps the sunniest city in the developed world Perth and have a rooftop solar unit. My unit delivers around 10 times more electricity in mid-summer than mid-winter. Yet I know my electricity consumption is almost the same. Clearly no saving in capital costs. I heard a few weeks back from someone who maintains one of the biggest wind farms here in WA that they were offline for a month because mice chewed through the control cables. Distributed energy infrastructure and wildlife don’t mix. @coldlynx, Willis For germany using hourly load and PV production figures over a year’s timespan it turned out that you’d need around 120 average supply-days of battery storage to make use of all your PV generation. Combining wind and PV turned out to be better (half the storage) but still prohibitive. they really do hate nuclear! 11¢<12¢. Solar power is still the best alternative for off-grid locations and activities, such as your boat, camping, holiday homes etc. I live in cloudy, rainy England but solar panels are appearing on roofs around the UK in increasing numbers. WOW!! Now that’s something that is worse then we thought !!!! @Bernie re wind power capacity factors – In Germany, it’s less: Turned out to be ~17% in 2009 and ~16% in 2010 (was more close to 20% some years ago). German win power incentives are designed to ‘take the pressure from the coast’ so less efficient inland locations receive some extra compensation. see this earlier post for cost including backup and load balancing The US has very low gas/oil prices – the UK /EU pay up to 4 times the US windpower costs: cost per kwh over 20 yrs £0.060 per kWh cost per kwh over 15 yrs £0.069 per kWh cost per kwh over 10 yrs £0.087 per kWh Willis, Thanks for bringing this article to our attention. I think a potentially more interesting story lurks in the comparison between wind and fossil fuels. A number of assumptions have been made that make wind look more competitive than it really is. One wonders if this was a deliberate politically motivated directive rather than a series of innocent assumptions… A number of people have noticed the very high capacity factor they have assumed relative to actual experience with wind.You also pointed out the maintenance costs appear to be low. The assumed 30-year life of all the plants also favors wind. Most experts would estimate the expected life of a wind plant at considerably less than 30 years — probably 20-25 years is more reasonable. On the other hand, most coal-fired plants are likely to be used for 40 or more years. Finally, buried in the fine print is an assumption that the discount rate is set 3 percentage points higher for fossil fuels than for renewables to account for GHG emissions! First off, a tax on CO2 (or its equivalent) would raise fuel costs, not the discount rate. Second, why throw this into a comparison of economic production costs. When all these corrections are made, one finds, not surprisingly, that the costs of wind exceeds the cost of fossil fuel generation by about the amount that wind has to be subsidized to remain competitive. As regards adding an amount for “backup capacity” that is, in a way, double counting the low capacity factor of wind, but not entirely. The short run intermittency of wind requires additional backup, not just its low average capacity use. In a sense, a discount should be applied to account for the lower quality of the wind power — or a cost added to make it of equivalent quality like you suggest. At the moment, I don’t think this cost would appear in the subsidies that wind generators need to receive to be competitive since the wind generators do not bear the costs that the low quality of their output imposes on the rest of the system. @coldlynx I see your point in spirit but not in practice. If the goal is to “get off the grid”, the initial investment of a diesel generator is a fraction of solar. There is still the ongoing cost of fuel, but that’s the sort of thing you can stockpile over time when prices are favorable instead of making a huge commitment up front. You can also make biodiesel pretty cheaply if you buddy up with some restaurant owners. :) Of course, there isn’t enough waste oil for everyone to do that. Generally I’m not a fan of any biomass fuel that requires fertile farm land to produce. High-lipid algae would be perfect if it could be grown in the desert or ocean and processed efficiently, but for now it makes solar look cheap. But really I have to question the value of getting off the grid (if the grid is available) because the grid is relatively cheap and convenient. I don’t see how completely disconnecting from the grid would make me more “independent”, especially if it means a significantly greater fraction of my 10-hour day must be spent on electricity in order to sustain that independence. I do have a generator and enough fuel to last through a couple of weeks of a power outage, which I consider the worst-case scenario. Any worse scenario (e.g. war), and I’m not going to hang out in my house whether it has solar panels or a backup generator. . Here is a weed covered 20 acre solar installation in green Germany that has only been in operation for about two years. Even with subsidies it apparently it isn’t worthwhile to do proper maintenance. I have not come across a wind project at less than 13c per kWh excluding back up costs, and that one was in an especially good place and had some hidden tax breaks. 20c is much more typical – if you are lucky. Most of the rest of the world of course has to pay about 3 times as much for natural gas at the moment compared with North America, so coal is coming in at lower cost for now elsewhere. Solar PV is fine for the householder who wants it – but please do not continue to expect everyone else to subsidise it and take any excess at ridiculously high prices in the euphemistically named feed-in tariffs – which are subsidies to the rich from the poor. Espen says: December 3, 2011 at 3:51 am Good post, Willis. Just one thing: when you complain about solar subsidies, don’t forget that fossil fuels are subsidized as well! ===================================================================== Really?? Can you name those “subsidies”?? Don’t confuse the tax code with “subsidies” I understand the 25-30 year life on wind installations is a pipe dream–more likely they wear out and self destruct in about 7 years, depending on quality of manufacturing, service utilization and attention to maintenance. It will obviously require some time before the final numbers are in, but from what I’ve seen so far, the service range in years is much closer to 7 than 25-30. In that case, wind power becomes even more prohibitively expensive. I detect a “green” agenda in the numbers put forth by the IEA, as Peter Hartley above suggests. Willis Eschenbach says: December 3, 2011 at 2:34 am Exactly, Willis! That’s consistent with the saying: “Anytime the government gets involved, it creates a surplus in one sector of the economy, a shortage in another, and the taxpayer makes up the difference.” Obviously, free market, non-governmental investment decisions aren’t hindered with such illogical and destructive utilization of capital ($). Willis – I have nothing but respect for your views – but you do seem to fail to raise the crucial flaw as far as solar is concerned. For twelve hours a day (averaged), every day of the year, anywhere on the planet – solar panels produce zero – because its dark. E.On is one of Germany’s biggest operators of land-based wind power generation (6+GW in nameplate capacity); in addition to their other generating facilities. Their areas of operations stretch from the North coast to the Alps; roughly 1000km. In their wind power report of 2005, they published a utilisation of about 16% and a conventional shadow capacity (spinning reserve) requirement of 98%. They state that wind power is limited due to its unpredictable behaviour; which requires a electricity separate grid from the main one, to reap the full potential. In terms of expanding capacity; they (IIRC) also stated that all the premium wind areas are being used, so further expansion would be to sites with poorer returns. So more wind power means more power stations capable of ramping up and down quickly to fill the holes in demand that wind power cannot supply. Oh, and lots more high-tension power lines. Still, it’s good for the Austrian economy and investors, who are expanding their pumped storage capacity; “buying” surplus electricity from (e.g.) wind power at a peppercorn to pump up the storage; and then sell back electricity at a premium, peak price when it’s desperately needed. It’s worth billions of Euros a year. They call that electricity “generation”, for some reason only comprehensible to politicians and marketing managers. So they’re all in favour of wind power in Germany. (And extremely peeved at the Czechs for having the temerity to expand their nuclear power capacity.) Wolfgang: Thanks for the link. There is apparently a lively debate in Denmark over the continued expansion of wind generated electricity with a target of 50%. I am looking to try to nail down the actual costs of wind. The Danes put a huge tax (beyond VAT) on household electricity consumption which turns the estimated generation and distribution costs from about average for the EU15 to the most expensive. It is unclear where that tax money goes. The one thing you all perhaps realize but are not saying: According to the greens, cheap, available electricity is bad. Expensive electricity is *good*. It is ther *goal*. I agree with Willis’ conclusion that, even if solar panels are free, the system costs will not drop dramatically. The recent plunge in solar PV panel prices is indicative of an industry whose profit margin has imploded. Some estimates predict over 95% of solar PV companies in China will be bankrupt and gone by end of 2012. Another interesting trend is the large number of planned large solar thermal projects that have switched to solar PV this year. The DOE Sunshot program has 2020 goals for solar PV of $0.5/W for panels, $0.4/W for balance of system, and $0.1/W for power electronics. This, of course, is based on peak DC power at beginning of life, and it does not include labor. Even if these goals are met, solar PV still requires backup energy storage, be it hydro, nuclear or fossil fuel in tanks or coal sheds. DirkH- “The trend is towards one inverter per module” I agree with this. I finally decided to install grid-tied solar PV for my home. The problem with one inverter per module today is that the inverter is now almost as expensive as the module. I just paid $0.80/W for the panels, but I see prices of $1/W for the micro-inverter vs $0.5/W or less for multi-panel inverters. This will change. It is clear to me that the inverter electronics could easily be integrated into the module envelope without any new engineering breakthroughs. I suspect this will start to appear (at a premium price) in the next few years. I decided to go DIY for my solar PV project after I received a contractor quote of $10/W for what I wanted to have installed. My DIY total cost (panels, inverters, wiring, support structure, feeder to house) will be about $1.15/W after the federal tax credit. Interestingly, this is already close to DOE’s 2020 goal, since my DIY also does not include the cost of labor :-) Our utility offers net metering integrated over one year. Any excess generation is reimbursed at the wholesale cost of bulk electricity, about $0.05/kWhr. I estimate a break-even time of about 6 years here in Florida. Unless we get bulldozed by a major hurricane, of course. Wolfgang: What is the translation of 37,3 Mrd kWh? I assumed it was a million kWh. Thanks. Will- You pure speculation about wind power, and small easy to maintain generators is a tactlessly bs argument. Just admit wind works. Old Construction Worker says, .” If you’re an old construction worker, you no doubt know that you grossly underestimate the number of union workers it takes to change a lightbulb. I’ve dealt extensively with unions, and can tell you that it will take at least two journeymen electricians and one apprentice to ensure that the power is off to the line the bulb is on. Then the supervisor, after moving a small stack of 2x4s out of the way, will call a fork lift operator to move a couple of pallets of construction materials that will be in the way of the light-bulb changing operation. He’ll then run around the site to find a laborer to help him carry the ladder to where it is needed. Then the six you describe will proceed to change the light bulb, after which the above-noted two journeymen electricians (and one apprentice) will undertake to restore power to the line feeding the light bulb. The fork lift operator will then return the pallets of materials to their earlier place while the laborer does likewise with the 2x4s. Then the supervisor retires to the office with the union stewart to discuss the grievance the union is filing over the supervisor’s moving of the 2x4s and assisting in carrying of the ladder, instead of having laborers perform all such duties as required by the labor agreement. The grievance will then be heard by the construction superintendent and the union’s district rep, and will go on to a panel of three arbitrators, which will eventually award the most senior laid-off laborer $100,000 in back pay since the company should have had enough laborers on the site to ensure that bargaining unit work was only performed by laborers union members. You obviously haven’t dealt much with unions if you think only six could possibly change a light bulb. “Check out the windfarms and count how many of the fans are not turning at any given time …” I did notice that both wind turbines at East Midland Airport were actually removed from their hubs and brought to the ground for maintenance within a few months of original installation. That is not cheap at all. The other consideration with wind is we don’t know what figure they are using for how much each actually generates as a proportion of its capacity. Remember real figures have come out not much more than half of manufacturers estimates, so if the latter were used to calculate these you will be close to doubling all costs. Contrary to the above, it IS extremely likely that gas prices will triple or quadruple if a large-scale shift away from coal to gas occurs, which is why it is not happening with the present low gas prices. Not to mention the fact that the EPA and states are going after fracking because it is push hydrocarbons and god knows what other chemicals into the water table and biosphere. And, don’t forget the peer-reviewed research covered in the past on WUWT that places a higher carbon footprint on gas than coal due to all the methane that leaks out into the atmosphere from the fracking operations. Willis, I think you are being generous to solar PVs when you assume that gas turbines can provide adequate back-up for when solar PVs can’t generate power (2c/kWh). If you have a look at the generation profile of a solar PV unit you will see that it can switch off and on almost instantaneously when a dark cloud passes overhead, or there is some other shadow. To respond in time you need the OC gas turbines running in stand-by mode (i.e. add the cost of burning gas) and battery/flywheel capacity to fill the gap until the gas turbine comes up to speed. If other sources can’t respond quickly enough then the voltage/frequency will drop and protection systems within the network will start load shedding. The only way we get away with it at the moment is because solar PV generation is a relatively small fraction of the total generating capacity, and there is enough spinning reserve from existing coal/gas/nuclear power stations to cover. So the idea that costs will come down once we get ‘scale efficiencies’ may actually work in reverse with the cost of back-up escalating. Also, the 2c/kWh capital cost for an OC gas turbine is probably an amortised cost assuming the turbine generates at a certain level over it’s expected lifetime, say 20% for an OC gas turbine peaking plant. If that now drops to say 10% to run in standby mode, then won’t the capital cost per kWh produced double to 4c/kWh? But then we have to consider that using gas turbines as back-up is only an interim measure as gas still generates CO2. So to comply with the expectation of people putting solar PVs on their roof (that they are generating no greenhouse gases) backup will need to be purely from battery/flywheel etc. Then we start talking somewhere around 10 times the cost of coal/gas/nuclear for intermittent souces such as solar and wind. For the purists, they may want to do this, but I object when they start increasing the power bills of people on lower incomes through the subsidies. Matthew W: it’s not my idea to call it subsidies: Another facet not fully considered in Willis’s excellent post is how over time and an increasing trend (Govt mandated of course) in the % of total power needs generated by solar and especially wind, the coast of operating traditional fossil fuel generation increases as density (revenue generated per coustomer) of their revenue stream decreases. The greater the percentage of wind and solar, the more expensive traditional sources become, California citizens, already suffering from this, will be taught again in spades if the current policies are carried out. Energy is the life blood of every economy. If one wishes to reduce the population growth of a third world country, then simply make energy cheap and abundant. Current policies, predicated on fear of large poulations, will also reduce poulation growth, but tragically this will be through famine and war. The leveling fails to take into account the lifetime of the generation system. The life of a wind farm is actually considerably less than the theoretical figures quoted here. There are already over 14,000 abandoned wind farms in the USA (google it). They are becoming an environmental hazard. Willis you alluded to the problem of “the trek you’ll have when you forget to bring the size #2 Torx head screwdriver …” but there will not be a long queue of engineers wanting to work on top of a 400 ft pylon in winter gales in the middle of a rough sea. Unfortunately, the politicians have been listening to salesmen not engineers.. These numerous sources of small random and intermittent “pops” of feedback to the grid are of no value to the power company by the simple nature of their randomness and intermittancy. The local grid has to supply 100% backup for these users while gaining absolutely no advantage from the electricity that they produce. In addition, NYSERDA a government organization, hands half of the cost of the installation to the individual user. So, as a regular electric customer, I am paying for both the initial installation subsidy and the increased cost per KWH that results from the mandated buyback of worthless power by the grid operator. I have no problem with off-grid use of individual windmills and solar panels but the grid connected homes with solar or wind do nothing but waste the materials used to build them and the energy used to back them up all at the cost of evveryone else in the neighborhood. Ny staste assessments to cover these costs add up to about $.015 /KWH and nearly $.05/therm for gas. Thank you greenies, renewables, and all of you eco-schemers making a buck on the backs of your neighbors. Some time ago, I downloaded some insolation data from a nearby agricultural research station and plotted the result. The graph shows total insolation (solar energy) onto a horizontal surface, over a month. Click here if no graph shows It illustrates the high variability of “supply” over a year. A picture that those selling solar probably don’t want their prospective customers to think or to be able to make a rational judgement. What the graph doesn’t show is the number of successive the dull days, when there isn’t enough sunshine to generate any worthwhile amount of electricity. To do that, I’d have to pick an arbitrary figure for “dull” like 1 standard deviation below average. And then you get a maximum of 2 weeks of dullness. But I’m not sure if those were the weeks in which the person responsible for cleaning the bird droppings off the heilostat was on vacation. Between 3 and 5 consecutive days of dullness are common. (I really need to get to grips with R because analysing a decade+ of daily weather data in a spreadsheet is a RPITA) I’m a big fan of solar panels for individual consumers. They add a lot of flexibility and resilience to your life-style. Large-scale utility stuff: Never saw the point to it. Among other issues, if you make it really large-scale you come up against real undoubted human-induced warming, though probably not on a huge scale. Cover a huge area in something black and intentionally non-reflective, which converts 15-20% of the incoming solar light to electricity, and what happens to the other 80-85% of the incoming solar energy? Can we say waste heat? The immediate surroundings of a large scale solar plant will be much hotter than they would be naturally. You can get higher efficiency in solar, but currently at a cost that means it only makes sense to use the cells in concentrator systems, which are okay for large-scale systems, but I wouldn’t want on my roof because of the heat issues. I wouldn’t write grid-tie solar electricity off completely though. The key is to get the efficiencies up inexpensively and the balance of system costs down. Pure silicon cells are approaching the theoretical max for the technology, but people keep finding ways to squeeze a few more percent out of it, and I suspect that someone will figure out a way to mix materials in an inexpensive way. Hit 30% efficiency and a lot of the balance of system costs go way down because you’re working with smaller panels for unit of electricity and that impacts land used, labor, etc. Barring a breakthrough that won’t happen before 2016, but long term higher efficiency is what you really want in solar. Nice article Willis. In general it looks fine. I’m a bit surprised that geothermal is so inexpensive. I assume that the costs are some sort of blend of existing sites which are highly optimized. I wonder if they include the disposal costs of high temperature brines such as those exploited in the Imperial Valley. I also wonder if the costs are lower than those that would actually be encountered if a serious attempt were made to tap geothermal resources on a large scale. I’m dubious that there are that many high grade sites available. I agree that onshore wind costs seem too low. It seems to me that as a practical matter the costs of any large number of turbines should include the costs of backup generation, pumped storage or some other sort of energy buffer. I also wonder whether solar costs don’t overestimate capital costs. I have trouble envisioning capital costs that high for PV installations in locations like the Mojave desert playas — no trees to clear, land lease should be cheap?, No need to construct service roads? I prowled around some in the paper and links but couldn’t find much information on exactly how the costs are derived. I imagine it is there somewhere. I suspect that the table that shows minimum and maximum costs is mostly latitude driven. The very first hydroelectric plant was installed in 1868. It is still operational with the original equipment. So can the figures be reworked to show say 10 years for a wind turbine and 100 years for a hydro scheme. The results will show a very different picture to that posted. Indeed the new chart should be publicized and the creators of the first bought to book. Willis, excellent essay. I agree that some of the numbers for “favored” sources seem optimistic. I used the EIA figures for actual electricity production for 2010 by source and the installed nameplate capacity for that year to arrive at the following capacity factors. Solar PV & Thermal 1299 Million kwh 987 Mw installed CF 0.15 Wind 94647 Milliom kwh 39516 Mw installed CF 0.27 If the installed nameplate capacity is for the end of the year, the real CF may be higher. I am assuming that these two sources are always feeding the grid when ever they can. In the meanwhile, much less expensive sources are held in standby or spinning reserve. I was wondering how the windfarms out West handled the high gusty winds earlier in the week. Anyone know. That might have an impact on estimating maintenance costs. In the emerging Ontario market a 10kW residential grid tied solar system can be fully installed for around $50k. That includes everything. Taking into consideration panel degradation, a system of that size will generate roughly 400,000kWh over the next 40 years. Inverters have 20 and 25 year warranties now too. That puts the capital energy costs around $0.125/kWh instead of the $0.20/kWh shown on the EIA graph. Move that same system to Los Angeles and the system can make up to 50% more energy in the same time. Capital cost/kWh is now down to $0.083/kWh. Now we are knocking on coal’s door. Rooftop solar provides the most energy during the heat of the day when air conditioners are pushing the grid to its limits and there are no transmission lines to pay for. It would stand to reason that larger systems on stores and factories can bring these numbers even lower with volumes of scale. And let’s not forget, these systems are not designed as a primary source of electricity that requires back-up. They are designed to reduce the use of finite resources such as coal and eliminate the need for peaking capacity. And yes, I agree we may be in a bit of a glut with respect to solar panel supply. That doesn’t mean we can’t make hay when the sun shines. MrC Espen says: December 3, 2011 at 6:35 am Matthew W: it’s not my idea to call it subsidies: ================================================================= Quoting someone that is wrong, won’t make it any more right. you says: December 3, 2011 at 6:25 am Will- You pure speculation about wind power, and small easy to maintain generators is a tactlessly bs argument. Just admit wind works. ============================================================= No one argues that wind doesn’t work. Wind does blow, that’s what it does. Trying to convert it into economically feasible energy isn’t what it does well. Willis, I agree with your comment on wind power. Even leftists/environmentalists (e.g. George Monbiot) in Europe are beginning to despise this “technology”. While in Hawaii (Big Island) last summer on vacation and heading toward the “Green Beach” we noticed that they had installed a new windmill farm on the Southern tip of the island. Next to it was the “old windmill farm” rusting away. To paraphrase Delingpole – the whole place looked like an abandoned War of the Worlds movie set. DOE is one of the most politically motivated departments in Washington. I’ll bet their figures for life cycle costs are way low. They probably are not accounting for decommissioning – but since it’s an “environmentally friendly’ idea it’s OK if it ruins some of the most beautiful scenery on earth and rots away in the sun. I’ll send picture of this mess if you want. Tom Kennedy @Willis .” Solar thermal only has to store process heat, not electricity. Also, I am suspicious that the Bechtel installation in S Calif is being used as the ‘cost’ (which is deliberately in the billions). Perhaps we should examine technologies that do not involve Big Green or Big Energy in the same manner that we consider Thorium in the face of opposition from Deliberately Big Uranium. Solar thermal includes space heating, correct? I am suspicious there is some way to better analyse the need for domestic energy, for example. A lot of domestic energy is thermal and does not need to be electricity between the sun and the applied heat. Jus’ thinkin’… Here is wind data for the month of December, 1999 measured at a now developed wind farm in New England. This was measured with calibrated sensors placed at 30 meters and 25 meters AGL. ( I will have to resurrect my old 80286 to provide the 12 months of data I have and apologize that the output is in mph and not metric M/S). Summary Wind Speed Statistics Mean Wind Speed: 11.1 mph SDEV of Wind Speeds: 6.4 mph Mean Turbulence Intensity: 0.191 Mean Energy Speed: 19.2 mph SDEV of Energy Distribution: 5.7 mph Mean Power Density: 156 Watts/m2 Max. 1 hour Wind Speed: 32.9 Assumed Average Air Density: 1.186 kg/m3 Energy Pattern Factor: 2.14 Weibull Shape Factor, k 1.80 Weibull Scale Factor, c 12.5 mph Speed Data Recovery 99.0 Turbulence Data Recovery 100.0 Direction Data Recovery 92.9 Note: This was one of the better months, the wind speed was recorded below 10 mph for a total of 290 hours. Summer average was about 8 mph. “Good post, Willis. Just one thing: when you complain about solar subsidies, don’t forget that fossil fuels are subsidized as well!” ===================================================================== “Really?? Can you name those “subsidies”?? Don’t confuse the tax code with “subsidies”” Willis, well said, excellent article. Typically when investigated, the so called subsidies are exactly as you say part of the tax code not the direct subsidies given to solar, wind, and biofuels. god. The tax code is by design complex to suit the political machines, and I have limited knowledge in this arena Keep in mind that with all these “breaks” the fossil fuel business pays huge taxes to the US treasury unlike the renewable fuels. Finally one needs to be aware that after income tax, lease sales and royalities from fossil fuels are the largest source of income to the US Treasury. What is the plan to replace that income source, tax the rich? Power production from wind turbines is very dependent on wind speed. Maps of winds in CA at 100m indicate very few places with adequate winds. Wind farms are constructed with one large turbine per 80 acres. That provides 1/3 to 1/2 mile spacing which reduces the risk of fratricide when a turbine fails and throws pieces weighing tons about 1/2 of a mile. 10% of CA electricity needs supplied by wind would require about 2500 square miles of area, considering the capacity factor. This much area with the required sustained winds doesn’t exist on land. Consequently, 10,000 or more turbines would have to be built at sea. Marine wind power costs much more than land-based wind power. That ignores the fact that levelized costs are all lies, skewed for political purposes. PV solar also requires significant water use to remove dust. Dust reduces power production to a degree (est. 10 to 30%) that washing must be done. The water cost is significant in infrastructure, maintenance, and consumable. Solar as an individual solution in theory makes sense, but only for people living on their own property. How does an apartment dweller make use of solar in NYC? Opportunity costs of solar and wind are high. Once you build those systems, the land is not very useful for many other purposes. Certainly, humans can’t live there anymore. In the future, people are going to need to live in the third dimension to avoid paving over farmland, forests, and wild places. We will need compact energy sources. Natural gas makes the most sense in the medium term, but nuclear is our solution for the long term. We won’t have any choice if we want to maintain our population. Who is volunteering to get off the planet? That’s the OWS-socialist fantasy, that someone else will have to die. As that article, and many others documenting the rapid obsolescence, decay, and abandonment of windfarms by “gone with the wind” ‘rupted operators, show, the real maintenance/replacement costs of renewables installations are likely to be disastrously higher than the official estimates. Like maybe an order of magnitude. One comment I’ve seen a few times from technicians is that it’s actually a rare bird(!) who can work at the heights of the wind-rotors — and the danger is vastly increased in windy conditions. Which–doh!–are the pre-selected norm where windfarms are located. Solar and windfarms are going to end up as hugely expensive salvage and clean-up operations. The gifts that keep on giving. Espen @ 6:35: The article you reference makes the same mistake many others do. It equates subsidies with tax credit. A subsidy is paid to someone to produce more. ie: ethanol, wind energy. A tax credit allows someone to deduct the cost of doing business. ie: mining, drilling, manufacturing. Wind companies can deduct the cost of the windmill but they then are paid a subsidy for producing the (intermittent) electricity. Big difference. Since we are stuck on the cost factor of wind generators right now, I recall reading recently, and had confirmed by my neighbor that works for a major electric utility; that the latest 2Mw wind generators require constant power to spin the motor in shipping and once installed still require power to spin the motor when there is no wind so that the bearings do not prematurely fail in it. So, if true, what does this really mean for the viability of large capacity wind generators. Also, do you know what happens when one of these generators “accidently” spins backwards? @Bernie This is from energinet.dk, danish TSO – they must export ~80% of their wind power cheaply to their scandivavian neighbours and need to buy it back expensively when their winds production fails. Here’s what it looked like in Germany:. The one in Malaga is definitely not me :-) Taking into consideration panel degradation, a system of that size will generate roughly 400,000kWh over the next 40 years. That is about $40,000 worth of power over 40 years at retail prices, for an investment of $50,000. You are losing $10,000 even before we consider what else you could have done with the $50,000. Add to this the fact that the system is not likely to last 40 years. More likely 20 years as the panels degrade, which would be only $20,000 worth of power. Which means solar power is able to turn your $50,000 into $20,000 in only 20 years. Now in Ontario, they have a FIT program that hides this true cost, by paying consumers 5 TIMES the retail market prices for electricity for solar power. So, while a $50,000 investment can return $100,000 over 20 years under FIT – unless a new government cancels the program – $80,000 of that amount comes from everyone else paying taxes. So, if everyone in Ontario was to do this, then everyone’s taxes in Ontario would have to go up $80,000 over 20 years to pay themselves the $80,000 difference between FIT and market. However, this won’t happen, because only the rich can afford the $50,000 investment, which means the poor in Ontario are paying the rich to install solar power. @Bernie Mrd is short for the german ‘Milliarde’=1e9 (Billion). Thus 37,3Mrd kWh =37.3*1e9*1e3 Wh=37.3e12 Wh. CF(2010) = 37.3e12 Wh / 8760 h / 27214.7e6 W = 0.156.. = ~16% Espen says: December 3, 2011 at 3:51 am Good post, Willis. Just one thing: when you complain about solar subsidies, don’t forget that fossil fuels are subsidized as well!——————- I keep hearing that from the current administration, but can’t for the life of me figure out what they are referring to. The own an interest in various wells. I pay tax on every penny which they produce, sometimes tax to federal, state, and county governments. I was able to to deduct initial drilling costs from the income, and deduct an annual depletion allowance of 15%, much in the way one deducts depreciation on a rental property. The difference between depletion and depreciation is that one claims depreciation even if the property appreciates in value, why the reserves in my wells actually do decline every year. Those deductions are the only “subsidies” I can think of. By comparison, last year one could put solar panels on a house and recover 80% of the cost through subsidies. Hiya Willis and Happy Holidays! Disagree, disagree… Solar (and wind, for that matter) are not intended to replace other fuel sources, only complement them. Best case scenario is 25%-30%.Within those design specifications for a smart grid, especially solar looks good. Others have noted that it is not just the module subject to price decreases due to innovation. In addition to inverter and other BOS components, there is the chance to lower prices considerably by reducing the bureaucracy associated with permitting and interconnection. Another factor to add in when comparing capital costs is (on the good side) we don’t know really how long modules will last, as they refuse to die so far and (on the bad side) inverters really need to improve their lifespans. Free fuel is compelling, especially to owners of large homes with pools who like A/C in summers. For now, that’s enough of a market to chase. As the price/performance ratio improves, the number of homes for which solar is appropriate will increase. It’s gonna be okay… I was wrong … the visual scan over the data didn’t pick up periods where data were being collected,, but there was low insolation. Here’s a graph showing the number of maximum consecutive days over the previous 30 days, when insolation was less than one standard deviation below the mean. (The blue line on the graph is meaningless; it only joins the data points so that you don’t go cross-eyed.) I’ve not plotted data for periods during which there were no data. Even for those without PV panels, solar hot water systems have been popular in Western Australia for well over 30 years because they work well for 8 to 10 months of the year. Solar hot water systems make more sense than PV because their energy is easy to store; and if you don’t have a house-load of teenagers, you may not need to add heat to the hot water that you use even after a spell of 3 to 5 dull days. The energy is stored in the water. Cheap. Space heating in the Perth metro area is seldom done by solar; not even supplementary. The cost of plant is so much higher than for other heating; unless you’re in a building with central airconditioning and heating. The ROI for even a small supplementary heating system in a domestic residence is negative especially if one includes the cost of plant maintenance (tanks, pumps, blowers, filters and valves). Matthew, barryjo: I’m not going to insist that it makes sense to call it a subsidy. But in any case it’s an advantage given to a specific industry, isn’t it? @Bernie Apologies for busting in… Mrd is short for Milliarden – a Milliard is equal to one thousand million. A term largely ignored by the English-speakers especially those leaning to the USA-version which adopted the term “billion” from the French for the same quantity, before the French went back to “milliard” because the “billion” was a bad idea in retrospect. USA-ian use of “billion” persisted and its (ab)use has spread along with the financial markets and dealing of the USA. Leading to much confusion. So I habitually avoid the term, referrring to it as thousand-million if I mean that, or million-million if I mean the other. Sometimes, I get lazy and call a milliard a milliard. Some people think that a milliard is 0.036 inches.:-) I don’t know about the rest of the world but, here in NY wind and solar are heavily subsidized while so called “fossil” fuels and hydro are heavily taxed. >> Espen says: December 3, 2011 at 6:35 am Matthew W: it’s not my idea to call it subsidies: << Then you need to look into the details rather than just read the headline of a news release. 1. A large part of the so-called subsidies to 'fossil fuels' is the ability to write off costs incurred in exploration/drilling and taxes and fees paid to foreign countries for the extracted oil. In any other industry these are considered normal tax write-offs as business expenses. Compare that to direct payments of tax dollars (e.g. Solyndra) and higher electrical costs required by the government for solar. 2. Another significant part of the subsidy is the government subsidizing heating fuel for the poor. This savings goes to the fuel companies but is passed on to the user. Compare that with subsidies for home solar installation, electric cars, and other green energy consumer subsidies which go mostly to the wealthy. 3. The real subsidy is measured in money per unit energy. Even if you claimed that business costs should be taxed and that the tax write-offs are therefore a subsidy the 'subsidy' per kilowatt hour are insignificant compared to those for solar and wind. Far from subsidizing fossil fuels governments tax them – think gasoline taxes. Every quanity of fossil fuel replaced by subsidized renewable energy must take into account the lost in revenues from taxes and added to the subsidies . In Europe these taxes are enormous. That is why gasoline costs 10 dollars a gallon. In California the state collects almost 3 billion dollars in gas taxes. Can you imagine what a mess California would be in if all cars were electric and subsidized. California is finding out that increased electricity costs is resulting in what the supporters of cap and trade described as ‘leakage’ – the lost of manufacturing jobs to other states or in the case of solar cell manufacturing – China. john says: December 3, 2011 at 7:16 am Here is wind data for the month of December, 1999 measured at a now developed wind farm in New England. This was measured with calibrated sensors placed at 30 meters and 25 meters AGL. Correction with apologies, I had the lower anemometer placed at 15 meters. Should be 30 and 15 meters. My results are consistent with the expected resource at almost all of the major wind farms located in Maine except for a couple located at the higher elevations in Maine’s western mountains. In Ontario the FIT program solves the problem with solar power not being able to supply power at night or when it is cloudy. The FIT program requires you to have two meters. One that measures solar production, and the other that measures mains consumption. You pay about $.10/kwh for mains power, and are paid about $.50/kwh for solar power. In other words, the power company will pay you 50 cents for the same thing they sell to you for 10 cents. Question: How much solar power in Ontario is the result of “leakage” from the mains to the PV systems? How many batteries banks backing solar arrays selling power at $.50 are actually being topped up from the mains at $.10 to increase efficiency? So long as you kept this within reason, how could it be detected? Given the financial incentive, isn’t large scale cheating inevitable? Solar panels with built in inverters would be a fantastic idea. No need for expensive wiring. Simply plug them into any outlet in the house, and as the sun shines, your power meter will slow down. Add enough panels and your meter will run backwards – unless of course you have one of the new smart meters. Willis: Does wind work as some claim? Have a look at the distribution graphs a the bottom of this page. The other wind sites are mapped as well — see the index. Note the median point — showing that 50% of the time the output is 17% of faceplate or less. Note the mode — the most common output is Zero (0). Wind power requires 100% backup by another reliable source — as does solar. Unless of course the sun can be made to shine throughout the full 24 hour day — all year long. Maybe in the “New Warmed World” the wind will blow all day — but we are not there yet. The wind dies out completely from time to time over the entire province of Ontario — see section 3.1 of the site. If people can solve these problems maybe renewables have a chance. Till then… Sunspot says: December 3, 2011 at 2:16 am Yes, and Denmark has the highest electricity rates ($0.43 kWh) of all developed nations, of all nations identified in this table: Yes, it is tough being green, but that is perfectly alright, as long as it is Joe Sixpack who is paying for it. Many thanks for making us aware of this very useful information. As an alternative to taking capital cost plus operating cost as the most important figure, one might take the view that the capital cost has less importance the more one thinks in the longer term. Presumably the capital cost is over the lifetime of a power plant, a few decades. On that time scale, we will probably be needing to substitute for large quantities of energy we now derive from oil. Noting that the replacements will need to come from different energy subsectors, and that corporate resistance will continue, society might decide that lack of resilience is a major externality. In this case, it could be instructive to look again only at the operating costs, with the idea that society might be well advised to make an additional investment in the future. In this light, and if one is skeptical of the operating cost of wind, then the low cost options that can come on line in sufficient quantity are nuclear and solar PV. I expect Willis will object to this interpretation; just saying that good data are food for many different thoughts. Espen says: December 3, 2011 at 6:35 am “Matthew W: it’s not my idea to call it subsidies:” Several comments. First that is an ancient article covering a period up to 2008. Does anyone have a clue as to how much subsidies to renewables have increase since 2008 under Obama? Second the Bloomberg article admits that the fossile fuel do not receive subsidies but rather cites tax credits as follows “The largest of the subsidies for fossil fuels in the report was a tax credit oil and natural gas companies can claim for paying royalties to other governments. The institute’s report finds that credit totaled $15.3 billion over the time period.” As indicated by others that is part of the tax code available to all industries, not a subsidy of the sort the Administration hands out to Solyndra, etc. Also keep in mind that the large US oil companies earn over 50% of their income from overseas investments/business. Does anyone believe that it would be beneficial for the US to tax foreign income returned to the US without giving some form of credit for foreign taxes paid? No foreign earned income would ever arrive on our shores if that became the tax policy, and many companies might just relocate overseas. Also because of lack of oil/gas investment opportunites in the US due to numerous restrictions the investments in oil/gas are already moving overseas. 10 oil rigs have already left the Gulf due to the administration restrictions post BP spill. Oil rigs are expensive and take time to build so that is lost production for an extended period, lease revenue loss, and royalities that are difficult to replace. ‘From the Article: “Also included in calculation of subsidies are the U.S. Strategic Petroleum Reserve — an emergency oil stockpile — and the Low-Income Home Energy Assistance Program, which helps some consumers pay for heating and cooling costs. ” Would any reasonable person count these as subsidies to the oil companies. I thought the Strategic Reserve was created for energy and national defense security, silly me. Finally to inculde Low income Energy Assistance.. as a subsidy to oil/coal indicates the agenda is to distort the facts rather than provide useful information. Nothing is credible in the Bloomburg article it is pure propaganda. I question the wind capital cost numbers: it is very unclear whether the EIA backs out the massive subsidies which every single wind electricity installation receives. If said costs are also modified via feed in tariffs or some other legislation like the California renewable energy mandate, this also distorts ‘true’ cost. A number of comments asserted that fossil fuels are being subsidized, while a good number of responses set those wrong assertions straight. Some years ago, when the price of gasoline was at about $0.75 a liter in Alberta (and about $2.40 a liter in Portugal), I ran across a study of world-wide gasoline prices at the pump, done by a Calgary consulting firm for the oil industry. Unfortunately, I lost the citation and have not been able to find the report anymore, in spite of intensively searching for it a number of times. The gist of the study report was that the cost of producing gasoline and moving it from the well-head to the gas tank at the pump, including all production costs and mark-ups at intermediate stages, excluding all royalties and tax revenues, was at that time between $0.23 and $0.26 a liter everywhere in the world. Taxes comprised the difference between that and the price at the pump. Here is a snapshot of the sources of energy generation in Alberta on one of the coldest days in Alberta, last winter: Source: “…although of the total Alberta generating capacity a full 5.8% is supposed to be derived from wind turbines, at 11:20 am only 0.022% or 2.2 hundredth of one percent were being generated from wind power.“ @WillR “Wind power requires 100% backup by another reliable source — as does solar.” Agreed as logical, however while you can’t store compressed air to blow at the windmill, you can easily store huge amounts of heat to run solar thermal. I have assumed that ‘storage’ is on all cases referring to storage of electricity in chemical or capacitative form. Suppose instead heat is stored and turns the generators 24/7? This is not far-fetched. One should look at the cheapest energy storage point in the total system and place the back-up there. Clearly wind is the worst candidate because it requires distributed on-site storage or perhaps a centralised pumped water storage (as found in Cape Town, for example). Not mentioned by anyone, that I noticed anyway, is the idea of “peak demand availability”. This factors in the availability of the equipment (on-line availability), the occurrence of peak demand (time of day, season), and the likelihood of the equipment actually having a source of power (wind or sun, for example). Here in Idaho, peak demand occurs in mid or late summer, when there is usually little wind, so the peak demand availability of windmill generated power is 5% of nameplate capacity during the time when it could actually do some good. That eliminates windmills as a viable power source, yet subsidized windfarms are being forced on the power company and the costs forced on the consumer and taxpayer in the name of saving the planet (and lining some pockets). Fortunately, we still have some of the lowest power costs in the country due to lots of hydro. Another topic: I noticed in the cost chart that nuclear is out of position. It was moved lower on the list than it should be. Nuclear is competitive with the latest fossil fuel technologies, has a great safety record, and emits no carcinogens during normal operations. Wolfgang Flamme says: December 3, 2011 at 7:34 am “Here’s what it looked like in Germany: ” Interesting; but that was when we still had all our nukes running… The gentleman that indicates that a $50,000 investment will produce 400,000 kWh over 40 years, for a cost per kWh of $.125 per kWh. I would think you might add at least two items to the cost: -the cost of the $50K outlay at 5%, would be about $2900 per year over 40 years. That is $116,000 total, or $.29 per kWh -$.02 for maintenance. The FIT program you talk about is driving the acceptance of solar and is intended to wean the industry off subsidies once the market matures. The slashing of solar subsidies are normal processes, not because they are non-cost effective or whatever, but because they both drive and respond to efficiencies found in the industry. Gotta love public-private partnerships. BTW, the FIT does not solve the problem of lack of solar energy at night. The grid tied nature of the system does that on its own. Currently Ontario has something like 80MW of small solar projects receiving $0.802/kWh (not $0.50 as you pointed out). That’s roughly 80 million kWh/year costing $64.16 million/year to the rate payers. When you consider that rate payers consume 151TWh/year the cost to the tax payers is an insignificant $0.000000424901/kWh. If the larger projects accounted for a 1000 times greater impact it would still only add $0.000424901/kWh. And FYI, for FIT projects, batteries are prohibited from being connected upstream of your solar meter. From a technical perspective it’s pretty hard to game this system, although I’m sure some will try. MrC @Crispin in Waterloo. I await your economic analysis of the feasibility of Wind/Solar storage. Any study I have seen simply shows that it is far more expensive than conventional energy. And to show my heart is in the right place I will give you a flying start for one type of renewable storage… Hint: It does not appear to be economic — I did the calculations for another forum… As for heat storage you might want to examine the efficiency calculations. I remain open to the possibilities that will be shown by your analysis… ;-) SoloPower is getting about $300 million in federal, state and local subsidies to open a new manufacturing plant in Portland Oregon. The local press has been cheerleading the news as a big jobs creator. But the press wasn’t so kind in San Jose having been jilted in favor of Portland. I pulled out the best quotes I call 10 red flags=fatal flaw. My favorite is #5. It’s the bulls eye- The founder accusing the firm’s major investors and board members of “unjust enrichment”. Isn’t that the whole point of these schemes? Never mind viability. Line your pockets while you can. Someone in some official capacity should be screaming bloody murder. Or at least whispering? After Solyndra, a 2nd Solar Energy Firm Is Scrutinized San Jose-based SoloPower uses the same risky technology and also received a multimillion-dollar federal loan guarantee By Aaron Glantz on October 15, 2011 – 12:03 p.m. PDT 1. Three weeks before Solyndra declared bankruptcy, the United States Department of Energy issued a $197 million loan guarantee to another Bay Area solar company that uses the same innovative, but risky, technology. 2. In its six-year existence, SoloPower has experienced internal discord — it paid a $20 million buyout to its founders — and has yet to turn a profit. 3.analysts say it is unclear whether SoloPower’s costs are lower than Solyndra’s. “They have not yet revealed anything on their costs of production,” 4. SoloPower has struggled to increase production to commercially viable levels. 5. Founder Talieh in a lawsuit, accused the firm’s major investors and board members of “breaches of fiduciary duties, abuse of control, gross mismanagement, waste of corporate assets and unjust enrichment.” 6. But Talieh and Basol were bought out for $20 million, and the lawsuit was settled. 7. At the Energy Department in Washington, Damien LaVera, a spokesman, said he had not been concerned by the developments. 8. “This application was approved after extensive review by the career professionals at the department 9. SoloPower is one of four privately held CIGS solar manufacturers based in San Jose. SoloPower and two others have admitted they had never turned a profit. A fourth company refused to say 10. Of the four, SoloPower is the only firm to be granted an Energy Department loan guarantee Doug says: December 3, 2011 at 7:53 am No doubt, some people will consider the taxes you pay on oil revenues to be a subsidy. Well, if those are subsidies, then everyone’s job is being subsidized as well, and everything that the governments taxes could be considered to be subsidized. Aside from that, the German Government actually provided tax incentives for those who installed solar panels on their homes or properties. The return on investment was quite good on that, and the scheme took off, mainly because excess power could be fed back into the grid, and compensation derived from that. However, not all of the government bureaucrats are stupid. Some of them soon figured out that the scheme was costing the government a lot of money of which it had increasingly no enough. Therefore the government soon began to tax the power delivered into the grid from any PV installations. Of course, that did not quite cover the costs of installing PV panels and other alternative energy schemes. No problem, the electricity rates were jacked up to more than $0.30 kWh to cover the shortfall. Perhaps some of our German friends can elaborate on where things stand now. Are the players in the solar craze sill as eager as they once were in both teams to waste their money on a losing proposition? Bruce Stewart says: December 3, 2011 at 8:36 am “Noting that the replacements will need to come from different energy subsectors, and that corporate resistance will continue, society might decide that lack of resilience is a major externality” Well, your “corporate resistance” goes both ways; as much as Big Coal doesn’t want to be dismantled, just as much do the wind power people lobby for more subsidies for their sector and fight other energy producers (they finance anti nuclear leaflets, for instance). In Germany, the PV and wind power companies have built up enormous lobbying organisations and their people constantly travel to Brussels to push their agenda on the rest of Europe. As for the “lack of resilience”, this is the true driving factor behind German governments’ push for renewables. Both types of governments use the CAGW “scientists” only as useful idiots. The real objective is to build up a more diversified energy base to reduce dependency on Russia who deliver most of our natural gas. These plays will be disrupted by more shale gas discoveries, though… I have used EIA reports off an on for years, these engineers do good work. It is usually so well documented that if you need to make your own adjustments they too will be reliable. Almost all energy sources have their place and utility in the grand scheme of things. It all depends on local conditions and circumstances. Some like solar are very much restricted and some like wind are simply a make work project for the repair people. It seems to me that to do anything other then gas, coal, atomic, hydro are based on non economic criteria. Downdraft says: December 3, 2011 at 9:20 am ….Another topic: I noticed in the cost chart that nuclear is out of position. It was moved lower on the list than it should be. Nuclear is competitive with the latest fossil fuel technologies, has a great safety record, and emits no carcinogens during normal operations. Something else needs to be put into the context, that is, for example, the cost of mining in terms of human lives lost. World-wide, about 100,000 men lost their lives mining coal during the last century. I am not aware of any lives lost in the mining of uranium. Your foresight could be worth a fortune! There’s an ETF (Electronically Traded Fund) that tracks the price of natural gal “futures”; I think its ticker symbol is UNG. Leverage up, catch the wave, and then you can lean back and light up that big cigar! One thing Governments and Green Companies don’t want to discuss is the “Human Costs” of Solar and Wind………………….the same old BS rhetoric bubbles to the surface like “Wind and Solar are clean and are replacing the dirty coal and fossil fueled energy systems in the world”. WRONG!!!………….for every single Kw of energy produced by the clean green wind and solar industry one Kw of gas generation is required when the wind doesn’t blow and the sun doesn’t shine!………….the particulates from gas fired generation are smaller than the ones emitted by coal and cannot be “scrubbed” with the latest technology. Asthma sufferers will notice this before anyone else. Clean Coal is a reality but not “politically correct” anymore. People’s health has been attacked by Wind Sound, and properties have been sadly devalued so that people living within the shadows of these Industrial complexes cannot sell and get the hell out! Of course the last thing “Greenies” give a damn about are Humans……..they also think there are too many of “us” on “Their” Mother Earth! I must confess that I cannot comment on the economics of PV energy generation, however my own solar powered hot water system was a really good investment. It is not subsidised by anyone else, I don’t export power, but it really made a major difference to my energy bills. Sometimes is easy to overlook how much energy we lose by heating how water and pouring it down the drain. I thought it would take 5 years to pay for itself, it took in fact more like 4. Everything now is clear profit. Renewable energy will never be able to compete with traditional fossil fuels on an economic basis, but it is not that simple. A previous poster commented on the mortality rates for coal mining in the last century. In addition to that were the many others who did not die, but who were severely disabled. The cost of care for such is an issue is rarely included in our calculations. Additional costs of the maintenance of families where the main earner is disabled is also discounted. It may be that such factors increase the economic cost of traditional fuels more than we care to admit. mike g says: December 3, 2011 at 6:33 am “ . . . push hydrocarbons and god knows what other chemicals into the water table . . .” Please educate yourself about the fracking issues! The “water table” is not involved. Hydrocarbons are the desired output, not input. Water and sand are the big inputs; . . . and gaur gum. You don’t need god to explain this. Willis, It will be three years from the time that the large RE (pv, concentrating solar, etc) contracts that PG&E, SCE, or San Diego have signed before the details of the contract will be know to the public- which will then allow the public to see what the true costs of purchasing the electrons (generation) will be. We can get a glimpse of the generation costs (which do not include transmission or distribution costs) for contracts that were signed off in 2009 from the following California Public Utilities Commission “ENERGY DIVISION RESOLUTION E-429 dated December 17, 2009″ located here- A key item for say NRG (the folks that have been buying up a lot of the large RE projects in CA) is the Time of Use factor that allow the generator to obtain premium prices for their power at peak times (which is when PV ‘s output hits it maximum capacity factor in both a day cycle and within the year). “The.” So for PG&E they will be paying NRG a price 2.20490 * the Market Price Referents for the output from the plant in the summer peak months. A PV facility starting up earlier this year with a 25 year contract with PG&E the Market Price Referent = 0.10442. By my math this means NRG will be getting 23.03 cents for each kwh they send to the grid at Super Peak times in the summer. I feel a bit sorry for the poor rate payer advocates at the CPUC as they are going to have to figure out how to allocate these costs to PG&E’s customers. Manfred says: December 3, 2011 at 2:58 am Sure they will lower all cost components … including all cost components for non-renewable energy resources. So it will not benefit solar. w. Here’s an idea: The standard royalty paid to the government for production of fossil fuels from public land is 12.5% of gross. The system is fair, in that those resources belong to “we the people” and we deserve our cut. Likewise, the sun and the wind belong to us all. Let’s collect a 12.5% royalty from tapping those sources as well. That would level the playing field and ensure we use the most cost effective source of energy. Rob L says: December 3, 2011 at 3:34 am Distribution costs in California are not seven to fifteen cents. I pay twelve cents per kWh all up. w. Several questions comes to mind. If I make any improvements to my house (new AC, etc), the value goes up and so do the real estate taxes, which are considerable in NJ where I live. If I install solar panels on my roof, do my real estate taxes go up? Shouldn’t that be part of the annual cost? If not, why not? Are solar panels they exempt? Will they be exempt down the road? Possibly if I sell the home at a higher price due to solar panels, will the new owner have to pay an increased tax bill, many towns base taxes on purchase price? One a crazy subsidy scheme is created by the folks in goverment, where does it stop? Espen says: December 3, 2011 at 3:51 am In the US there’s a tax advantage if you drill for oil in the US. Other than that I don’t know of much in the way of subsidies for fossil fuels. What subsidies are you speaking of? w. JC says: December 3, 2011 at 6:08 am Couldn’t agree more. The idea that inexpensive energy is bad for the world is perhaps the most anti-human of all the “green” ideas. w. Feel free to learn from your Northern Neighbors… Canada — if you have heard of it…. 3. Watch out for financings, as usual. Investment bankers strike when the iron is hot and they can sell deals. This in and of itself implies solar stocks were peaking when all the financings and IPOs came out. Basically, it’s the age-old rule. When everyone else is buying, maybe you shouldn’t. 4. Watch out for industries that rely on government support. Far better to invest in a sector that is self-funding without subsidies. The government has enough hidden surprises for you on the personal tax level, so don’t double-up on your government risk by buying companies in need of government help to meet their business plans. Pay attention to lesson 4. Cheers! mike g says: December 3, 2011 at 6:33 am Not true. If the price of gas goes up like that people won’t use it. The EPA thinks that sunshine and pure air need regulation. I doubt greatly that the loons are enough to stop fracking. And no, it doesn’t “push hydrocarbons … into the water table”, that’s your fantasy. Do your homework, crying “Wolf!” doesn’t work around here. Yeah, people care so much about carbon footprints … in addition, that research has already been superseded. w. Walter H. Schneider says: December 3, 2011 at 9:34 am “Perhaps some of our German friends can elaborate on where things stand now. Are the players in the solar craze sill as eager as they once were in both teams to waste their money on a losing proposition?” Yes. The left parties call the cross subsidy, which increases electricity prices by about 17%, unjust, though, and want a progressive income tax instead to finance the renewable madness. The FDP (classical liberals) is making noises that call for a limit on the new capacity of installations in a year. Guaranteed 20 year tariff for new installations is cut by about 20% each year, following the cost curve of PV. Still all gung ho for Solar in cloudy Germany. thebiggreenlie, WRONG!! For every kWh of energy demand, we can supply that with either a renewable source or fossil/nuclear. While base load nuclear power generation is more of a “light it and go” nature, fossil fuel generation can be modulated. Less use of fossils means less finite fuel consumed, less wear and tear on the equipment and less pollution like SO2, mercury, arsenic, cadmium, lead, selenium and uranium being emitted. (I do not consider CO2 to be a pollutant). Fossil fuels don’t “back up” renewables. They’d be needed anyway. We just need less of them as renewables are doing a bit of the work. MrC coldlynx Would you find the same joy by growing all of your own food, and with growing/producing your own textiles? Perhaps you would enjoy building your own wagon from wood to be pulled by your own horses and grinding your own lenses for glasses. Heck, let’s get totally off the grid and not by pv panels, just burn what biomass you can grow. Getting off the grid sounds great, but improved efficiency – lower cost (your labor and resources) – comes from larger scales. Why do you dislike power companies? Rural electrification is still recent enough that some remember the time before – it was not easy. The power provided to all through electricity by these companies enables our civilization to continue to grow at the fantastic rates we’ve seen. I do NOT want to go back to having to survive on what I alone can produce. Incidentally, here’s a point VMartin made here on Nov. 20: Espen says: December 3, 2011 at 6:35 am Thanks, espen. Your citation says in. Further down in the article someone says: What he said. Finally, even according to the citation’s ludicrous figures, the so-called “subsidies” to fossil fuel companies amount to $72.5 billion on the 93% of the US energy the fossil fuel companies produce, while subsidies to renewables (which produce ~ 5% of total US energy) amounted to $29 billion. Even their bogus figures say that per unit of energy, renewables are getting about eight times the subsidy as fossil fuels … and real figures would make the disparity even worse. w. If you want the concise condensed version refer to my comment on the most recent DOE comparative electrical generation cost report. This is one my staple documents. One might also take a look at biomass which is just another way of harvesting solar energy and it’s comparable in cost to advanced coal, wind, and nuclear. Here’s the table and this picture is worth a thousand words, with far more data, all in one nice spreadsheet format and quite unlike the typically colorful but comparatively empty Figure 1 chosen by Willis. you says: December 3, 2011 at 6:25 am Assuming that was meant for me: 1. My name is Willis. 2. Yes, wind works, it works very well. Wing generated electricity also work, just not for long and at a huge cost and not very well. 3. The numbers I gave for wind power are backed by citations. What you have just put forward is “pure speculation”. 4. “Small, easy to maintain generators”? You ever done any generator maintenance? Try doing it hanging off the top of a hundred foot tower and tell me how easy it is … w. Hi Willis, As soon as I read the very low alleged costs for wind power I knew they were wrong. The stated capacity factor of 34% is far too high – low 20’s is routine. But the Substitution Factor, as indicated by E.On Nets Wind Report 2005 (emailed to you) is the killer. Wind power typically requires more than 90% conventional backup. Wind power, which varies as the cube (^3) of wind speed, often provides power when it is absolutely worthless, and furthermore can destabilize the grid due to its wild short-term fluctuations. Finally, if wind power was any good, it would not need huge, life-of-project subsidies. I expect that we will ultimately conclude that wind power is utterly worthless, and perhaps even generates less valuable, peak-load power than it consumes, on a life-of-project basis. Regards, Allan _________________________________________________________________________ Willis Eschenbach says: December 3, 2011 at 1:40 am Walter H. Schneider says: December 3, 2011 at 1:30 am … Then there is the question of whether the cost estimates for wind power are based on the theoretical maximum capacity rating of the turbines or on the real generating capacity of around 22 – 24 percent of rated generating capacity.. Here’s a good widget for energy consumption and production: I cannot believe that on-land wind is as cheap to build and run as indicated in the chart. Why are there 14,000 defunct turbines in the US? Maintenance is greater than is indicated, broken parts cannot be recycled and are long term pollution. What is not indicated with wind, on or offshore, is that it functions terribly, only when the wind blows, and even has to be turned off when the wind is too great, at about the speed where the turbines become efficient but the gears cannot take it. Also, wind makes power at inconvenient times and has a number of related health issues to humans and animals. Turbines in the UK have been paid to turn off when making energy when it is not needed; they make money for doing nothing! There are wind farms in California that have to turn off for 5 months of the year to allow bird migrations. There is a cost to having all of this investment, either not making wind during the correct times of day, times of year, or not at all, as with the defunct mills. This all raises the cost of on-land wind to the same stupid level as offshore. Willis Eschenbach says: December 3, 2011 at 11:14. ========================================================= Stuck your foot in your mouth again, Willis. Do you know the difference between a tax credit and a tax deduction? Evidently not. Or you didn’t read carefully enough. Good grief. This is simple tax stuff but like just about everything else it’s way beyond your skill set. Willis says “Try doing it hanging off the top of a hundred foot tower and tell me how easy it is …” In my neck of the wind swept West — towers are a bit taller, the climb harder, and the sway more noticable. “. . . The towers are 221 feet (67 m) tall, and each rotor is 129 feet (39 m) long, . . .” I would only add that working on any large structure entails its own work-related issues. The issue is the number of facilities and their ability to be profitable and pay for maintenance and repair. Regarding home systems (aka off grid), many home owners don’t know how to turn the water off entering their homes. Good luck with power systems! Then you have to know how to turn the outgoing power off also, or your system becomes a danger to utility workers during a grid outage/repair event. >>On a totally separate issue, I suspect that the maintenance >>costs for wind power are underestimated in the report, that in >>fact they are higher than the EIA folks assume. I fly over these wind carpets daily, and from my experience, about 15% are out of action at any one time. That is not only a high maintenance cost, but a significant loss of production too…. . Very interesting article Willis. It sparked some good commentary. One thing though, if you consider something occurring in the early 1950’s as being the advent, then I guess Hydraulic Fracturing qualifies. Somehow the idea that fracing is a new phenomena seems to have ingratiated itself into the critique. I know it is a lot scarier if it is new and untried, but it is old and commonplace. I have heard it said that there have been over a million frac jobs in Oklahoma alone since its inception. I think what you mean is shale gas which as an exploited phenomena is relatively recent. I will say that if your intention is to stop shale gas drilling than outlawing fracing would certainly bring it to a screeching halt. I can’t help but think that this is the motivation behind some of the scare stories. Further to my post above, about 15% of wind turbines being out of action. Please note that these are NEW wind-farm installations, less than 5 years old – so it will be interesting to see the situation in 30 years time. . MrCannuckistan says: December 3, 2011 at 7:09 am The “emerging Ontario market” is supported by an enormous subsidy. You folks are paying 42¢ per kWh to the producers of solar power, Wikipedia sez you are discussing jacking this to 80.2¢ per kWh … and you want to boast about this? That’s the biggest subsidy on solar I’ve heard of. Costs in their study are levelized over a 30 year period. You need to compare apples to apples. That jacks your numbers by a third. So already we’re up to 17¢ per kWh. In addition (as you point out) the inverters will likely need replacing. That puts us at maybe 18¢ per kWh. To that we need to add a couple cents for operations and maintenance, there’s no free lunch. Call it 20¢ per kWh all up. That’s three times the cost gas at 7¢ per kWh … it only works because of the subsidies. Same objections apply. It’s still twice the cost of conventional gas power. Same problem. Why would a store want to cool itself with solar power that costs twice or three times what conventional power costs? (Transmission costs as a part of levelized solar costs are only about a half a cent.) The problem is you want other people to pay for your hay habit. Other ratepayers are forced to subsidize your power preference, a preference which doesn’t work without huge inputs of money. This raises the cost of power to everyone, particularly including the poor folks that cannot afford to install a fifty thousand dollar yuppie fantasy. It is a hugely regressive tax. It raises energy costs for everyone, and I think that is a pernicious plan which does not benefit anyone. w. Espen says: December 3, 2011 at 8:05 am Matthew, barryjo: I’m not going to insist that it makes sense to call it a subsidy. But in any case it’s an advantage given to a specific industry, isn’t it? =================================================================== It is not at all about “making sense”. At the risk of being called pedantic, the tax code is not a subsidy. That’s it. It’s very simple and very clear. Willis, The thing that bugs me the most is that the power company is selling me the same electronsl over,and over again. 3600 times a minute these little charged wave/particles see-saw through the wires of my house. At least that used to be the case. I have a vague memory that the frequency of my electricity would no longer be closely controlled allowing it wander somewhat about the 60.00 Hz mark. Two results: Now I get new electrons over time AND many of my appliance clocks are no longer keeping good time. Cheers First comment, anyone who uses solar east of the Mississippi and North of the Mason/Dixon is never going to see a payoff on that investment. I know of nowhere that solar is a “good investment” where there is no subsidy for solar. I will disagree with this. My company sells mobile solar and solar/wind power systems that are economically attractive today in remote locations and in certain applications in telecommunications, military power, and other areas where the cost of transportation for propane, diesel, or other fossil fuels is prohibitive. The U.S. military is paying $29 per kw/Hr (yes, twenty nine dollars per kilowatt hour) in Afghanistan. Willis, everyone: thanks for clearing this up. I was assuming that the Bloomberg article made at least some sense, e.g. that the tax deductions in question were unique to this industry. Matthew W is right. And subsidies only reduce taxes. In the case of oil companies, the taxes paid are substantial. Subsidies only reduce their tax liability by a relatively small amount. And “subsidies” like the oil depletion allowance are no different in principle than depreciation on a commercial building. Sal Minella says: December 3, 2011 at 6:38 am. +++++++++++++++++++++++++++++++++++++++++++++++++++++++++ Yes – that is the kind of cost I found when researching solar years ago, and even though solar prices have fallen, add batteries for emergency power or off grid applications and costs soar. Add snow, and dirt and clouds and capacity factors and you better not want to run much more than your fridge and deepfreeze. So I have a propane powered 16 kw backup generator as I live in a remote area with wells and a water to water heat pump to run. And I have to clean snow off my solar panels for my fencers and pond aerators regularly. Today it is snowing and overcast so they are all running on battery back up – till the batteries go flat. Get a dose of reality. The only way Solar panels seem to be “economic” is where there are big federal and local subsidies. Look at the following payback calculation from a PV company for North Carolina. Without subsidies the payback is close to 30 years (at 20 cents a kW) . About the same period as the replacement period for the equipment. Add batteries and a more competitive power rate and PV’s are totally out of the picture. ——————————————————– Promotion from PV supplier follows: Details: 2KW (or 2 Kilowatt) is a small starter system that would provide electricity with a value of approximately $45 per month. This sized system would cost approximately $14,000 before taking advantage of tax credits. The federal tax credits are 30% or approximately $4,200. The NC State tax credits are 35% or approximately $4,900. The bottom line cost, after credits comes to approximately $5,000. Note: The price of a 2KW system is approximately $7 per watt. As the size of the system increases, the price per watt would decrease. For example, a 5KW system would cost approximately $6 per watt. The typical 2 KW Photovoltaic (PV) system will include: 10 collector panels of 16 square feet each or 160 square feet of collector panels. Inverter, wiring, tie-ins, etc., for a complete turn-key installation. All electric work performed by a NC licensed electrician. Value / Payback: Most NC customers sell their solar electricity back to the grid, and get a total of about 20 cents per KWH (Kilowatt Hour). This represents 5 cents for the electricity, and 15 cents for the “Renewable Energy Credits”. A 2KW system produces (on average) about 7.5 KWH per day or 225 KWH per month. At 20 cents per KWH this comes to $45 per month, yielding a simple payback of 9.25 years. NOTE: In North Carolina, payments for “Renewable Energy Credits” come from NC Green Power, and payment rates can vary. Check with NC Greenpower before starting a project to confirm current rates. I’ve done my own calculations for the cost of solar simply using the published cost a project, then amortizing that number over the expected lifetime of 25 yrs to get a monthly payment. Then I take the kW or MW rating of the solar project and multiply by hours in a month time 15%-20% output. I’ve never calculated a number less than 38 cents per kWhr. And that assumes the average lifespan of a unit is 25 yrs, when in actuality the average lifespan is likely much less. Very interesting. Your post neglected the fact that the levelized cost calculation for GHG producing generators such as coal-fired without CCS includes a capital cost penalty, which has a similar impact to a $15/tonne carbon tax, to provide equivalency to low GHG producing technologies, without this coal-fired power would be cheaper still and the contrast with wind and solar would be even greater. Following from the DOE document: a $15 per metric ton of carbon dioxide (CO2) emissions fee when investing in a new coal plant without CCS,” Another one bites the dust: “Range Fuels Inc., a cellulosic ethanol company backed by as much as $156 million in U.S. loans and grants from President George W. Bush’s administration, is being forced by the government to liquidate its only factory after failing to produce the fuel.” A 30 year life expectancy drastically over estimates the amortized cost of Gas, Coal, Nuclear, and Hydro. If properly maintained, Hydro lasts until something breaks it (like a major flood event destroying the dam). Gas, Coal and Nuclear have many facilities that have been in use for 60+ years already, and despite reference I have seen that large boilers need something close to a rebuild after 30 years hard use, the rest of the facility is already built and the rebuild often increases the capacity of the plant. Wind is almost uselessly intermittent without some type of high capacity storage. PV while more predictable, has significant times when it produces nothing. The two most common forms of bio mass are wood chips and methane. American coal plants are designed for crusher run coal, with little to no modification they can burn wood chips only issues are lower burn temp and more ash per KWH. I have seen methane captured by some municipal sewage systems and some power plants get theirs from landfill gas. Willis writes “So panels are cheap, but not likely to get a lot cheaper in the near future, the market is still correcting. Finally, as I said above, panel costs are a small part of the whole equation.” What you are seeing there is the base product (ie the cells themselves) have dropped in price but not the whole framing/inverting thing. IMO these too will drop substantially over time once we settle on standardised ways of doing things and there are plenty of installers around to meet demand. No one renewable will solve all the problems and so we shouldn’t look to use solar for all the answers. IMO nuclear energy will also form a significant part of the solution in the medium term. The big worry about letting the market purely dictate what energy sources we use is that if/when hubbert kicks in and production falls away with non-renewable like oil, then we’re hit with the double whammy of spending additional resources to convert to another technology during a period where those resources are becoming increasiningly difficult to find. That, IMO, is the recipe for producing the worst in man. Willis, why let Coldlynx slide this by, “That give in Sunny states a annual output of up to 2 kWh electricity.” An annual output of up to 2 kWh or US $0.30 at high rates, is bad payoff. If he meant 2kW power or2kW per hour averaged over the year, ok but for economics of location for rate of return on investment. Some input from Germany: This is the Official Wind Power Monitoring Program website, formerly run by ISET Kassel, now part of Fraunhofer Research: As for costs: – spinning reserve backup costs are estimated to be around 2.5ct€/kWh_windpower – however no official estimate is available – neccessary grid enforcements are TSO’s responsibility (big issue right now since grid congestions become increasingly frequent) – hardship clauses in case of grid congestions (TSO’s emergency management, wind power shedding) For the city of Hamburg, Vattenfall recently announced a municipal cooperation project where they intend to use wind power generation surplus for district heating assistance… Most people in this “Energy sources” debate seem to be missing a simple point. If the householder spends money on Renewable source? That’s money, now, not wondering the economic system. It’s a diversion of resources to renewables. Try this for a thought experiment?… For those that want to install renewables as their main power source? Send extra taxes to the government and have them organise it for you, see how that turns out. Because that is what is actually happening. It’s not ALL about which techinology is or isn’t more viable than the other. It’s also about efficient use of resources. Most of the western world power generation has now moved to the private sector. Why? because the private sector uses resources more efficiently. The proposal for renewables requires State involvment. Why?. The private sector wont touch it on any meaningful scale because it’s in-efficient. OK, so you want renewables? No Problem. Write the government a letter stating you’ll pay extra taxes, or alternativley get them to cut the funding to your Schools, Hospitals, Roads and Infrastructure and see how all that works out for you. Double foot-in-mouth for you David. Your criticism of a minor confusion of “deduction” and “credit” adds nothing to the topic at hand. Come back when you have something to contribute. TimTheToolMan doesn’t seem to have much faith in the free market. As fossil fuel prices rise, alternate energy sources naturally begin to take over as the price becomes competitive. There is no need for government subsidies, in fact such subsidies are harmful to society. Productive funds are diverted to less efficient energy sources, raising energy prices and leaving less money for everything else. If there were no subsidies for wind and solar power, they would account for almost no energy production. By giving heavy subsidies to the Solyndras of the economy, while absolutely refusing to allow the extraction of the copious new fossil fuel reserves available off shore, the government is fulfilling President-elect Obama’s promise that “electricity costs will necessarily skyrocket.” How is this situation good for the country? Where does the 1c for Hydro fuel cost come from? I didn’t realize rain and snow cost anything. Thank you. I too found it a trollish distraction. The self-styled architects of our future energy scold us on the topic of “sustainability”, yet not a single one of their economic plans is sustainable. The most important component is cost and they act as if their good intentions can cause expensive power supplies to be superior to those that are far less expensive. Well, there are only so many jelly beans in the jar. The family that struggles to pay their $200 electric bill will not be able to pay one when it rises to $500, no matter how clean, sustainable or in harmony with Gaia new power sources may be. But what is really going on when people who advocated solar power for years suddenly oppose a solar power plant because of how it might disturb a newly discovered endangered species? That they really didn’t yearn for solar power after all. Demands for solar power were really a means to an end- to end our prosperity and national power. And I am afraid this is really the end goal of too many advocates of green power. Just a couple of notes – The US EIA uses average costs. Not many people live in a place called ‘average’. The cost of coal varies by 400% to 500% in the US(The further you are from Gillette,Wyoming the pricier it gets) Natural gas varies by 50% – the further you are from the Henry Hub the pricier it gets. @ buckwheat. This is about the associated financial derivatives and has nothing to do with energy. They do not even have to sell electricity. The following is applicable to solar and was brought to my attention awhile back. You might want to take a moment to read the SEC filing. This is from a company known as First Wind. They have numerous shell and shelf companies including UPC renewables etc. This is from SEC filings to present. (they have not yet gone public). . At a BBQ last week I found out that a couple had invested $38,000 in solar panels and energy. They took out a loan! Then said so far they had got back $160.00 for one house that they spent $12,000 putting in solar. And 1200 dollars for the larger one, that they spent $26,000 dollars. They are alarmists of course. Now I am wondering, doing our sums, how much will the interest be on those loans? In comparison to what they get back in electricity credits? Solar panels are only guaranteed for five years. And of course they may have taken out a loan for only a few years, but the rest of us, are subsidizing their electricity. Not fair. To illustrate this point the pending First Wind Holdings Inc. SEC S-1 and S-1A application for an IPO readily admits that producing electricity it is not necessary to be profitable. Back in my home town of Aberfeldy, in bonny Scotland, they are cutting down all of the trees I planted in the hills to the south of the town 30 years ago to builb a bloody windfarm?!?!?!?!?! Must be money in it for someone. ,,,builb…? Hey – no – it’s OK. It works. I’ve made a new verb: “to builb a windfarm”. Can anyone tell me, that if you have credits from either solar or wind turbines, is it taxable. Jimmy Haigh at 3.54 pm. Google Newburgh in Fife, Scotland. Of course there are people in it for the money, especially if they own the land the turbines are being built. It will force up the electricity costs for those who don’t own the turbines, as found out in California. While the others are earning subsidies and also credits for themselves. In someways, we are privatising energy sources and those involved are getting no returns worth the expenditure and input. My first time using tags so I hope I did this right…. The cost of ownership over the useful life of the product is a better oranges to oranges comparison. Just because your nuclear plant needs a major overhaul costing billions of dollars at the thirty year mark doesn’t mean you can impose its limitations on my choice of technology. Speaking of costs, the CBO says the average nuclear plant built in America is 207% over budget (table 2-1). Triple! I think I mentioned inverter warranties, implying that they may not need replacing. In the past four years inverter warranties have increased almost linearly. Now a 25 year warranty is leading the way. My car is warranted for 3-5 years but lasts 10-15 or more. My TV is warranted for a year or two but lasts 10-20. Lastly, there are virtually no O&M costs on a typical residential system so the 2¢ per kWh you allocate for that should cover the new inverter(s) if they are even needed. I did mention that solar was “knocking on coal’s door”, but either way coal or gas are both finite resources that will run out one day. I hope we have a solution ready for when we do. Which leads me to the subsidies you mention. No one says it doesn’t work without the subsidies, yet. As mentioned before, the idea of the subsidies is to give the industry the push it needs to become self sufficient. You obviously missed my post on the impact to rate payers. The overall impact is insignificant but the long term benefits to society will be profound. When those first, most highly subsidized contracts start to expire we will look back at an industry that will have made huge technological advances that brought them well below the cost of centrally generated fossil fuel electricity. And since you mentioned it, how about we talk about the insurance subsidy for the nuclear industry. By underwriting the liability of nuclear we, the rate payers, are not having to pay the billions a nuclear plant would have to pass on if they had to carry the appropriate level of liability insurance. Most of the people I talk to say that residential solar panels are covered under your existing homeowner’s policy at no extra cost. MrC Mr C Tou write: .” isn’t that what your solar power provides….subsidised power for a relatively small number of days. Or else I am completely missing the point of your content-free post Everett writes “TimTheToolMan doesn’t seem to have much faith in the free market. As fossil fuel prices rise, alternate energy sources naturally begin to take over as the price becomes competitive.” Thats fine if there is no R&D requred and no long lead times with infrastructural changes. It’d be fine if we were talking about the new iPod or something that is a want and not a need. The fact is there has been massive subsidy into renewal energy sources and we’re still not at the point where we can properly transition to them although we’re a lot closer than if we’d spent nothing to this point.. And when the supply is of a resource that is necessary for survival then combined with human nature, that becomes a major problem. “Lawrence Poe says: December 3, 2011 at 6:28 am Old Construction Worker says, “This is why the government wants wind and solar. It puts puts more people to work even though it is an inefficient use of capital.” Sir, I stand corrected. LOL I completely disagree with this statement. The complete lifecycle environmental impacts of solar are completely ignored from the mining of the materials required to manufacture them, the building of them, and what are we going to do in 10-20 years time when they all start failing? There is going to be a constant supply of dead solar panels. They can’t be recycled, the cells must be thrown away. Besides, at best a solar installation provides benefit only during the day. In most of the country they provide little benefit in winter because the sun angle is too low and the days are too short. You can cheerlead all you want, solar DOES make sense in very isolated locations where mains problem is not available or to provide domestic lighting or some other insignificant use of power but they will never provide power at industrial scale. Neither will windmills. How many windmills does it take to power one single electric arc steel mill 24x7x365? Heck, you can’t power ANYTHING reliably 24x7x365 with wind or with solar without even more environmental impact from such things as batteries for storage in off-peak generation times. I’m sorry, solar is absolutely no substitute. We have two sources of power for industry: nuclear and fossil and at some point in the next century, we will be down to one, nuclear. Solar on any significant scale is a pipedream and it always will be except for certain niche applications. It is an absolute waste of money for base power generation. It is not environmentally friendly and it never pays back without a lot of subsidy. Same with wind. It just doesn’t work and there is no possible way for it to work. Anyone who has lived in an area that might be cloudy for weeks at a time can tell you that. The costs are simply not worth the benefit unless you assume that not using conventionally generated power is some sort of benefit and so the entire exercise in large scale wind and solar comes down to convincing one’s self that conventional energy use is somehow “bad”. It isn’t. Put solar in the desert and it destroys habitat. Put it on rooftops and it is destroyed by hail, wind, etc. Sure, it might work great in Phoenix but it makes a lot less sense in places with more than 200 days of heavy cloud cover per year (Portland, OR; Elkins, WV; Binghamton NY; Kalispell, MT and thousands of other communities). And he is correct, it is a very regressive tax. I would call for the elimination of all subsidies for wind and solar in order to spur development. Subsidies hamper increases in efficiency by making the current level of efficiency cost competitive so there is no need to improve. But in any case, it is a stupid idea except for corner cases or as a hobby project. It makes no economic sense whatsoever. Lawrence and Old Construction worker. The facts are these, for every 1 job in new energy supplies, ie. Solar and Wind, has cost 2.2 jobs in conventional energy supplies. Some jobs security?. Wind turbine manufacturers are closing, solar panels are in dire trouble too, and people are losing jobs. It is cheaper to replace a wind turbine than mend it. And solar panels sometimes don’t last the 25 year distance they are supposed to do. After 5 years, you are up for the cost of repairing them. And there again, they are replaced at hire cost, not repaired. They still need back up electricity, those that signed up for subsidies before a certain date, get 60 cents per kwatt or equivalent. Who are the biggest wind turbine and solar panel manufacturers China! P.S. The gripe I have is some solar panel customers, say those that complain are not seeing the big picture. We are investing to save the environment and avoid climate change (oh yeah), not the money we get back. Just as well, because the cost will outweigh all environmental benefits, as they won’t do a thing to change the climate. Dream on solar guardians. Without subsidies the solar and wind industry will not exist.). The variation in my systems output in the winter months has been a lot larger (17% to 25%) then my experience in the summer. This winter time variation in output is directly related to how much snow I get at my home an how cloudy/rainy the month has been. My system keeps my marginal usage from PG&E to Tier 1 or 2 prices most of the time. For PG&E their current residential prices (e-1 rate schedule- ) are- Tier 1= 12.2 cents/kwh, Tier 2= 13.9, Tier 3= 29.3.0 and tier 4 and 5= 33 .3 cents/kwh. Essentially my system offsets what would of been my Tier 3, 4 and 5 usage from PG&E (Tier 5 prices were 50 cents a kwh a few years back by the way). At the moment the baseline usage quantities (the amount kwh that you can use from PG&E at Tier 1 prices) is set up so that most individuals fall into Tier 3 usage so their marginal costs are 30 cents. Earlier this year PG&E reduced the baseline quantities by 10%. TTTM says: “When demand exists and supply necessarily drops off then there is a bit of a problem.” I agree with that. But in this case the problem is artificially created. There are enormous reserves of coal, and huge reserves of oil under the continental shelf, but the Administration will not even allow exploration, much less extraction in the red zones. Why not?? China is drilling only thirty miles off our coast, partnered with Cuba. The U.S. has the safest drilling record despite occasional accidents. Who are we going to sue, if and when Chinese drilling causes a spill? The government is distorting the free market to appease the enviro crowd by drastically limiting the available supply, and the result of that bad attitude is becoming disastrous to the average person. The price of a barrel of oil is at least double what it would be if the supply wasn’t artificially limited. The problem is government, not the free market. Two years ago I taught a science camp on power generation. I had to be honest with the students after I saw the numbers. I calculated the cost of outfitting my home with a modest array of solar panels. Back of the envelope calculations showed that it would take nearly 50 years to break even without subsidies. Of course by that time part (most?) of the system would also require replacement at additional cost. Here is the poster I designed to compare the scale of power generation from various sources @ How many solar cells would I need in order to provide all of the electricity that my house needs? See diogenes, Based on your characterization of my post, yes, you are completely missing the point. You must have gone to the socialist school. Solar power provides a relatively predictable amount of power over the course of a year. If it didn’t, no one would build, buy or finance solar. The peaking/transmission that I mentioned in my above post is primarily built to service the summer air conditioning demand, something that solar can do quite well. With enough installed solar we can eliminate the peaking plant and also slow our use of finite fossil fuels. MrC Crispin in Waterloo says: December 3, 2011 at 7:14 am Solar thermal can store process heat, although not all that much, a day or so. It still needs backup. No, in this context “solar thermal” just means the generation of electricity by using the heat of the sun rather than photovoltaics. w. At what point should investment priority be given to improving the efficiency of DC household appliances and energy storage, bringing the total cost for decentralized solar power down. At a glance, using Willis’ numbers it would seem a 10% improvement in compressors and fan performance would make more sense then squeezing panels a bit more. Additionally, a standard for household DC, receptacles, etc. would help. McCunnackstan, do you believe that millions of years of coal resources will be used up when? And interesting point though. Hydro electric won’t be used up. Nuclear won’t be either. But I thought that clean or green energy was being encouraged, even if it is more expensive, especially to heat houses in cold weather, to save the planet? Cut down carbon emissions, that are controlling our climate? And then tax it and sell carbon credits so some manufacturers will have to pay to produce electricity, mine, drive cars. Mixed messages here I think. If it works well why change it. If it doesn’t work too well, why invest in a white elephant. If another ice age does come in the next 50 years, we will have to invent heat banks to store warmth, and use gas to heat our food, or a barbeque type oven. TimTheToolMan says: December 3, 2011 at 5:04 pm .” That’s a straw man. Energy demand is flexible. Tom Fuller says: December 3, 2011 at 7:56 am Tom, always good to hear from you. 1. I don’t know what you mean by “replace” vs. “complement”. If I cook with electricity five days a week and in a solar oven two days a week, does the solar oven “replace” gas, or does it “complement” gas? Your distinction is unclear. 2. Why on earth would you want to replace fuel costing 6¢ / kWh with fuel costing ~ 22¢ / kWh, whether all of it or just 25-30%? Call me back when solar gets competitive with gas combined cycle due to “price decreases due to innovation”. Until then it’s just wishful thinking. Panels refuse to die, but they definitely degrade, disconnect, and get dirty, and often much quicker than we’d like. Inverters refuse to degrade, but they definitely die. Willis’s Rule says that everything costs more and takes longer, even when you take Willis’s Rule into account … “Free fuel” means nothing when the equipment to harvest it costs 22¢ / kWh. More to the point, I love the idea of “the number of homes for which solar is appropriate will increase. At present that number would be ZERO if we used actual generation costs … except of course that PGE has jacked the electricity rates so high that for some of the rich solar is appropriate (at least if “appropriate” is defined as “makes money for the rich”). I fail to see how subsidizing the rich homeowner is “appropriate” in any sense of the word … Not true and already demonstrably not true. We’re already (in CA) paying huge subsidies so the wealthy liberals can assuage their green guilt by using solar to power the AC on their McMansions. That might be OK for you, but for me, it sucks … w. A few things that have not been mentioned. Where are most of the solar panels manufactured? China. Why not in the US? Because it is too expensive to make them here. Why? Because it takes a lot of power to manufacture these things, and the power in the US is too expensive because we have mandated use of expensive solar power. China does not mandate solar power, or even much clean power, thus, they can use lots of dirty cheap coal power to make these panels. If solar were so good, the Chinese would manufacture the panels, then install them in China, then use the power the panels produce to manufacture more panels. They do not, coal is the only power cheap and available enough to allow them to make these things. Thus we see that China knows that solar power is no good for actual power generation, only for milking other countries out of money due to mandates and subsidies. I imagine, now that the subsidies are drying up, that they will stop manufacturing these things, they certainly don’t need them except to sell to rubes like us. Now that they have milked us dry they will move on to something else. Second, it takes a lot of power to manufacture solar panels, How many years must that solar panel be around before it makes enough power that there is a net power gain? To really see how good power generation is, you would need to subtract the power it takes to make the generating plant from the amount of power it will produce over it’s lifetime. If one does that, one will see that solar produces a lot less power than we think it does, since it starts off by using a lot of power to even be manufactured. Considering that, just how much actual net, surplus power does solar actually produce anyway? Perhaps the reason we think it produces net, surplus power is only because we have shipped the power costs of making these things overseas were we do not have to use power to make power. If we did, we would probably still be running a net power deficit. What we are really doing here is importing dirty coal power from China, and converting it into solar power. However, we think we are being “green” because we do not have to actually see all that dirty smoke. Thus, we see that having China manufacture our solar powers has the following effects: It makes more total pollution worldwide, due to China needing many cheap dirty coal plants to have enough power to make these things. We lose jobs, due to our more expensive power causing companies to move to China since it is too expensive to manufacture anything here. When we lose jobs, the tax base goes down, and now we cannot afford the subsidies. And since it takes a lot of power to manufacture solar panels, we don’t actually gain much net power anyway. Wolfgang Flamme says: December 3, 2011 at 1:55 pm “Some input from Germany: This is the Official Wind Power Monitoring Program website, formerly run by ISET Kassel, now part of Fraunhofer Research:” Thanks Wolfgang – one diagram is VERY interesting. “Specific investment costs” (EUR/kW of installed nameplate capacity) are EXPLODING – probably the easy terrain is populated now and new installations become more and more expensive as they need to build in more difficult terrain… Bruce Stewart says: December 3, 2011 at 8:36 am I don’t understand how oil fits into your argument. Almost no electricity is made from oil. That’s why it doesn’t show up at all in Figure 1. Additionally, there is an ongoing shift already happening from oil to natural gas. This is because we found out that gas is available in huge volumes by splitting underground rock (fracking). This gas will easily cover us for more than “a few decades”. This is perhaps the best energy news of the last 50 years, lots of clean-burning fuel. This good news has been studiously ignored by many people, including the President and hosts of folks who think they are “green”. You desperately need to go start and run a business for a while, even if it is just a lemonade stand. Your idea that we can compare options by looking “only at the operating costs” is … well, it is certainly contrary to my advice as an experienced businessman. That’s the beauty of the net present value aspect of levelized costs, it lets us compare capital and operating costs across the board. Why on earth do we need a “low cost option” to natural gas? Natural gas IS the low cost option. If you want a second option, then nuclear. But for a third option, look up at the top. If you are agitated about carbon, coal with carbon capture and storage comes in at 10¢ a kilowatt-hour … why would I want your cockamamie scheme at twice that cost when a) we have coal and b) we can burn it cleanly? Wouldn’t be any fun if we all agreed, would it? Thanks, w. @Hultquist Google Frack aquifer I’m not normally a greenie weenie but it doesn’t take a rocket scientist to see that these processes are adverse to the environment. Much better to scrape coal off the surface and burn it than to pump chemicals under our acquifers to break the rocks above to free the gas. That is what fracking it. Willis, I have a number of serious problems with the numbers and methodology that the EIA is dishing out here. Any competent estimate of costs is given in ranges, not fixed amounts. For example, two hydraulic sites in Ontario, as built without cost of capital. One produces power for approximately 2.5 cents, and the other if built will produce for a minimum of 18 cents even though it could put out at least 10 times as much power as the first. Hydraulic is totally dependent upon site conditions, water volume and head, let alone transmission distance. To give an absolute number is a very bad joke. The same problem exists for fossil fuels as well. Generation costs will vary enormously with transportation distances. Again, these always have to be done with cost ranges, not absolute values. Just think how expensive it would be getting Powder River coal east of the Appalachians. Now there’s also a bit of jiggery pokery going on with their wind costs as well. These will be based o.n the wind maps of the US geological survey. I know this because Canada uses essentially the same methodology. These wind estimates in now way reflect the actual capacity factors a wind turbine will produce; the government assessments cause these to be inflated by at least 50 per cent from what will be actually achieved. You are quite right about maintenance costs being underestimated. The world’s largest off-shore wind fleet at Horn’s Reef has been a nightmare of huge cost overruns and very high outage rates. Salt water tends to eat fragile things like gear boxes and inverters. Now as to solar, I’m not interested in theoretical calculations. I’m much more interested in actual contract amounts. In the case of Ontario, its electricity ratepayers will be paying 80 cents, that’s EIGHT-ZERO for every kWh of power out of its 1000 MW facility being built by Samsung. And there’s no T&D cost in this, as Hydro One has to forgo the connection charge. Now as to nuclear, these costs can be inflated easily by the government of the day. Quite simply, a large proportion of nuclear costs are in the hands of government regulators, and they can drive it as high as they like. For example, German utilities collected a plant decommissioning fund. In the late 1990s, the German government levied an asset tax of about 10%, not on earnings but on the entire value of the fund. Decommisioning requirements were not reduced, so the utilities had to pick this up with rate hikes. And the governments of both Britain and the United States have been pilfering decommissioning funds for years and dumping them into general government revenue. Neither of these governments intends to repay nuclear utilities for the funds they have misappropriated, and so all of this will be absorbed as higher generation costs. Finally there’s a great deal of misinformation about Denmark and its energy supply. If you look at their energy profile they claim that 30 per cent of their generating capacity is in “Renewables”. Then you have to look at what’s included as renewable generation. Yes, there’s wind. But the largest source of renewable generation in Denmark comes from their highly efficient waste incineration systems. Wind is barely over half Denmark’s renewable capacity and much less than 10% of total energy generated. Also, Denmark’s wind fleet is predominantly in Jutland on the west side of its transmission system. This means that primarily its wind supply goes to Germany. Denmark’s coal fired stations have maintained or increased their output every year regardless of wind turbine construction. I’ll say that again for the slow learners in the back; not one single Danish coal fired station has been closed or reduced its output as a result of wind turbine construction. @Willis Contrary to the above, it IS extremely likely that gas prices will triple or quadruple if a large-scale shift away from coal to gas occurs, which is why it is not happening with the present low gas prices. Not true. If the price of gas goes up like that people won’t use it. —- Already happened at least once in the past ten years and they did stop using as much of it. The extra 3 cents/KW fuel adjustment charge from that price bubble has only recently fallen off my bill. Willis DOE’s SunShot program seeks to fund R&D to cut the costs of solar thermal power from 21 c/kWh to 6.6 c/kWh to be directly competitive with baseload power. NREL provides SAM Solar Advisory Model to help calculate the efficiencies and costs for solar systems. Willis, with respect to your post at 8:48, you are quite right. At $3.50, gas is the cheapest choice, but only now. In 2004-6 it was up around $7-8. This is the problem with gas; its extreme price volatility. And that volatility itself is a cost. During the winter of 2005-6, gas turbine operators in NEPOOL and NYPA were shutting down despite rise in winter demand, because the regulated price of electricity sank below the cost of the fuel. This is a typical feature of the breakup of the vertically integrated utilities in the 1990s, and there’s no getting away from this system vulnerability now. You are right to point out the vast amounts of gas extractable with new technology. However, by and large we do not know what the extraction costs of this gas will be on a commercial scale. Will it be as cheap as conventional gas? Highly unlikely. None of the above means that gas is not a highly useful fuel for electricity generation. What it means is that gas can provide supply within specific constraints, and those constraints largely exclude it from base load generation. The unpleasant fact for far too many energy technology advocates is that every electricity system works best on a very large scale where you have a wide mix of sources supplying the system. Nuclear, coal and hydro for base and intermediate loads, and gas for peaking demand. Solar and wind provide nothing for system reliability on either a regional or a local basis and are only built because of government fiat. If a utility actually had a real choice based purely on economics and reliability, neither of them would ever be built in any quantity however small. Actually, pv modules are fully recyclable. There are even industries set up to do just that. (See here.) While it is true that solar only works during the day, in winter they can produce up to 55% of their summer peak. There are also plenty of places in the country have enough sunshine to make solar work. (See here.) I agree. No one has ever said solar would power industry. No one that I know has ever made that claim. Solar is not base load, not now and probably not ever. Its use has always been for the peak of the day when demand is greatest. The power that industry needs is more available when locally installed solar eases demand on the grid. And I hope the sun is still shining in the next century too. Solar panels are typically designed to handle a 1” hail stone at terminal velocity and local building codes ensure that panels don’t blow off roofs. Any freak occurrences of bigger hail or stronger winds and you’re probably on the phone to your insurance company already. I’m guessing that’s a reason solar isn’t more popular in the hurricane/tornado belts. We don’t sell a lot of snowmobiles in Miami either. Competition in industry drives the efficiency you talk about. If Company A can do it cheaper than Company B then they get the sale. This drives costs down. With a Feed-in Tariff, the subsidies are re-evaluated periodically to make sure this is happening and align itself with the long term goals of grid parity. As mentioned in a previous post, in the first two years of FIT, Ontario has reduced installed costs by 43%. MrC MrCannuckistan says: December 3, 2011 at 9:28 am (Edit). Subsidy free??? What part of subsidy involved in the 42¢ / kWh payment don’t you understand? You be sure to let us know when the market matures, my optimistic Northern friend. Some of us who have been exposed to your “once the market matures” nonsense ever since the 1970s you mention above are understandably doubtful of the market maturing much any time soon. I understand that your panels will last longer than thirty years. The problem is that combined cycle gas power plants also will last longer than thirty years. So we have to pick a common period to compare them apples to apples. That period is 30 years. So for $50,000 capital costs you get 300,000 kWh over the next 30 years, which is about 17¢ / kWh. Add 2¢ for running costs, that’s 19¢ per kWh. If you are waiting for the market to mature, at 19¢ / kWh it will be a very long wait. w. To date there is not been one documented problem with fracking and aquifers. That is confirmed by the EPA. Not one. Ever. The fracking goes on thousands of feet below the aquifers. When they develop solar and wind that can regularly stand up to tennis ball sized hail, we might have something. you says: December 3, 2011 at 6:25 am Will- You pure speculation about wind power, and small easy to maintain generators is a tactlessly bs argument. Just admit wind works. ______________________________________ A picture is worth a thousand words: From Natural News, which is NOT a right wing enclave by any means. So the big corporate wienies walk away with the $$$ and leave the mess behind to be cleaned up by the land owner or the tax payer. Windmills were nothing but a wealth transfer mechanism. Transferring dollars from the 99% to the 1% as Occupy Wall Street would say. Too bad they are too blind to see this occurring right under their noses. Now how about we fines the SOBs for every bat or bird killed and collect some of our money back…. OH that is right no use suing a bankrupt corporation. Gentlemen The Advanced Coal with CCS figure looks wildly optimistic. If you look the DOE figures you can see the COE increase for CCS is ilisted as 24% – not the 35% the DOE is officially targeting for future technologies. Placing this in perspective the chart has advanced coal with levelized cost of 109.7 $/Mwh and advance coal with CCS at 136.5 $/Mwh. Calculating… (136.5 – 109.7)/109.7 = a 24% increase in the COE… not the 147 $/Mwh a 35% increase in the COE would produce. For the official DOE targets see here at:). It’s also worth noting the DOE’s CCS numbers do not a include the cost of transport, injection, and long term monitoring. I see another serious problem with the DOE figures. If you look at the capacity factors they have listed for both coal plants and coal plants w/ CCS listed as having 85% factors. I see two major problems with this assumption: 1) A well run base-load coal plant can hit a capacity factor of 90% for the first 20 years of its life. In others words until it’s fully depreciated. A 85% capacity factor is reasonable figure for an older plant with maintenance and dispatch issues. But you wouldn’t make a build decision to build a base-load unit with such a low figure. 2) It is not realistic to assume a CCS system with sensitive downstream Carbon Capture equipment, no CO2 storage capacity, and a need to have continuous flow into a pipeline is going to be capable of maintaining a 85% capacity factor. Hence, the 85% capacity factor the DOE is using for plants with CCS looks wildly optimistic. This is one of a number of major reasons utilities are not prepared to endorse CCS as a commercial technology… just too many unknowns Bottom line… these are DOE fantasy estimates. Looks to me like the DOE/EPA are proceeding with their current dishonest presentation of CCS as being both commercially available and cheaper than it actually is. Kforestcat When it is 102 degrees F in the shade in Texas? And it is an hour before dinnertime? What do you do – just SHUT down a whole society? Stop living? Where do you reside anyway – La La land? . Dave Springer says: December 3, 2011 at 11:29 am Dave, I’ve worked as a tax accountant. A tax credit is deducted directly from your taxes. A tax deduction is deducted from your income. I am well aware of the difference between a credit and a deduction. I passed the freakin’ U.S. Government test to be an enrolled agent tax preparer. I also was quite accurate in what I said about my own situation. If I pay a tax overseas, it is subtracted from my US taxes, not taken as a deduction against my income. Having lived overseas for about a third of my life, I know more than a little about these questions. Finally, why do you have to be so snarly and ugly and nasty? All it does is make you look like a vicious, vindictive little man. I know a hell of a lot about overseas taxes. I used to do the taxes for the US Consul in the Solomon Islands. Give your assumptions about me a rest, Dave. I’m neither a bad guy nor a fool. Yes, I may be wrong, I definitely have been more than once … but that’s no excuse for you to be a dick about it. And when after your unpleasant abuse it turns out I’m right, as in this case, then you look really, really foolish. w. Gail, 14,000 wind turbines have been abandoned because government has never placed decommissioning requirements on renewables they way they exist for any other form of generation. If, God help you, you’re a farmer who allowed a bunch of these for rent money on your property, you’re going to be stuck with the abandoned hulks at some time in the future. Dennis Ray Wingo says: December 3, 2011 at 12:28 pm Sorry for my lack of clarity, Dennis. I meant grid-connected solar systems. As I mentioned in the head post, there’s lots of places off the grid where solar makes sense. w. Gail, wind does not work. Well, not on a large scale. I might be able to light my house with it if I get the batteries and the charging controller and have enough wind averaged over the day but I have lived in places in the US where we went without a breath of a breeze for a week or more in the middle of a blazing hot summer (mid Atlantic states with a stagnant Bermuda high just sitting there with nearly 100 degree heat and humidity nearly as high). No wind day or night for a week or two. Wind is fine as a novelty but it can not be relied on for national infrastructure. Unlike conventional power where a major wind or ice storm might take out your distribution for a week or three, a major ice, wind, or hail storm in this case takes out your generation capacity that takes a lot longer to replace and is more expensive. Are we to replace a regions entire generation capacity every time a hurricane or hail storm hits? That is just dumb. Grey lensman says: December 3, 2011 at 7:06 am The very first hydroelectric plant was installed in 1868. It is still operational with the original equipment…… _______________________________ Bingo! I have the feeling the numbers for nuclear are way off too especially if thorium is brought up to commercial use. Bills in Congress dealing with thorium: Somewhere in my archives I have an ad from BC Hydro from the 1980s which, to paraphrase, essentially says to wait about 20 years for the market to mature before they would consider solar panels. Now we are supposed to wait 20 years for the market to mature, that would put BC Hydro’s original estimate at about 50 years for the market to mature. I will try to find that BC Hydro ad and scan it and post a link to it. This will either take minutes or months to accomplish as my archives remind me of the warehouse scene at the end of the Indiana Jones movie. Someone mentioned 102 F in the shade, what about 44 C in the shade (Australia Penrith years ago) at a dog show, some dogs dropped dead from heat exhaustion, and people were jumping into the nearby river fully clothed with their dogs too. Lucky the sharks didn’t come up that far. Anyone mention Bermuda, lived there couldn’t do without air conditioners in the bedrooms. Humidity plus temps phhewww. Had to keep a covered heater in the walk ins to stop leather mildewing up. Even the sea was warm, like stepping into a tepid bath, or a tropical aquarium. With the little tropical fish swimming around you it was great. I’m no water baby but I could stay in the sea for hours and not get hypothermia, different from Oz and UK. crosspatch says: December 3, 2011 at 7:29 pm “. . . aquifers . . .” The ‘not normally greenie weenie’ (although he is one today) started with god and the water table. Then returned with rocket scientist and “acquifers” and including “breaking the rocks above” something to free the gas. If I ask what that ‘something’ is, I suppose I’ll hear about brain surgeons and stratigraphy. Learning is such a slow process. If he ever gets the words and facts straightened out, we can move on to discussion. If I understand it correctly, this study leaves room for too many fudge factors and assumptions, which are easily manipulated to sway the final result. For instance, as many have stated, the land-based wind costs don’t pass the “smell test” because the study’s predicted costs are significantly lower than the known costs associated with wind farms that are currently operating. I understand that they’re looking at “new” generation sources, but why not simply use the range of ACTUAL operating and capital costs for plants in each category, as many readers have done in their comments? The apparently understated costs for land-based wind leaves one wondering what other figures may be skewed. This tends to discredit the whole study. As a EE & MS, having worked in the energy & alternative energy industry (engr., operations & maintenance) for 40 yrs the thought of harnessing wind and PV energy has always attracted me… probably just for the satisfaction of harnessing new technology & initiating the learning curve involved. However each time I get really excited, I redo my homework, due dilligence & cost analysis, and I find that I really can’t breakeven on a 4-5 Kw PV grid-tie system even after 20 yrs…. (Colorado Spgs, CO). I’d really love to put a system in, & I’ve got the perfect place to do so, but I just don’t have the extra $20-$30K hanging around that I can play with and essentially throw away… so to speak. But having an extensive applied engineering background I always like to look at competitive renewable power supply ideas & proposals to see how practical they are & how well they hold up to realistic economic evaluations. I recently did a critique of Colorado Springs Utilities Renewable Energy Sources (RES) power generation plans which call for an initial 10% RES utilization eventually graduating to a 20% RES power supply. My comments as listed below cover both wind and PV solar systems in general and identify many of the problems that so many WUWT readers have also noticed and already pointed out. Resource citation & info: EWEA: European Wind Energy Association, 5-Volume Study: Publication Wind Energy: The Facts, Part 3, Wind Energy Economics Based on detailed audits and studies of EU community wind farms there are several key areas of concern which I have highlighted below. Each of these needs to be carefully studied and evaluated with detailed sensitivity analysis of each area of concern: 1. Developing a wind farm with the intent of establishing base-loaded power availability is extremely difficult, given the variable nature of wind and the high probability that wind will not be available during peak power usage periods. Research has shown that wind farms only capture about 20-30% of the available wind (kinetic) energy and generate on average about 20-22% of their full load capacity. Therefore to obtain an average of 1 MW of power, one must provide approximately 5 MW of wind generation. At a rough cost of $2.3 MM/MW this turns out to be a very costly proposition. This is what the UK and the rest of the EU are currently realizing. 2. Wind farms (regardless of size) will not reduce the base-load requirement for hydrocarbon based power generation facilities due to the intermittent nature of wind. One can’t simply capture and store the wind or sunlight for use on demand. In contrast, with hydrocarbon based power generation, high density potential energy is readily available for conversion into kinetic energy on demand. This is the reason why conventional power plants are so popular. They are relatively small, easy to build and operate, incredibly practical and will always be required until high density power storage becomes practical and demonstrable. 3. Wind energy (and solar energy for that matter) are often based on a 20-30 year economic life. Conventional multi-fueled hydrocarbon based power generation facilities are based on a 40-60 year economic life. The short economic life of renewable energy sources is especially problematic, since maintenance costs approach replacement costs often before the first 20 years of the life-cycle period. 4. Operation and maintenance costs for wind farms can approach 25 – 30% of the overall power generation costs, especially with intermittent or zero wind flow conditions. 5. Renewal energy sources (RES) such as wind or photo-voltaic (PV-solar) are not necessarily benign to the environment. As research and analysis has shown, substantial quantities of birds are very susceptible to being killed. Videos of these turbine blade bird kills have been displayed on YouTube. 6. From a power security and reliability standpoint, both wind and PV-solar farms are susceptible to damage from severe weather due to natural occurring events such as storms or heavy weather, or from human sabotage. Remote, low density power sites are very susceptible to being rendered useless by either of these possibilities. 7. At the present time, the learning curve for RES (wind) is very steep since design parameters for wind generators, e.g., size, basic construction, blade design, required maintenance and operation are evolving very quickly…similar to the design of early personal computers. Overall costs of turbine towers, generators and maintenance are decreasing as operational experience with existing wind farms increases. The maturing of optimum engineering designs can take 30 years or more. I would caution a quick rush into procuring wind generators based on capital incentives, rather than careful analysis of overall system costs. A more cautious approach would be to set up a pilot plant with only a few generators on an intended wind farm site and then use the operations and maintenance costs, data and experience gained to carefully plan a commercial size facility, if it is economically justified. This avoids the risk of placing all your eggs in one basket. This is common practice for petro-chemical plants utilizing new or non-developed technologies. 8. There is considerable interest in developing RES to reduce CSU’s carbon footprint. This is very noble, but not very cost effective regardless of who, how, what or why it is mandated. It is generally understood that if we were to totally de-industrialize the US economy, it will not measurably reduce the world’s carbon footprint. The world’s engineering and scientific bodies have a considerable amount to learn in order to gain a more intelligent understanding of how CO2 affects our environment. The preponderance of data so far suggests that the world is generally better off in a more CO2 rich environment as opposed to the opposite. I definitely hope for my sake and yours that they don’t outlaw CO2… it could be a fatal decision for all of us. We don’t NEED thorium. We can use conventional power. PLEASE read “Smarter Use of Nuclear Waste”. You build a facility with two conventional plants and one reprocessing facility. After the initial fuel load, the only thing you ever bring into the facility is natural uranium 238, no enriched uranium ever moves again. No nuclear fuel ever leaves the plant. The waste decays to background in a few hundred years instead of tens of thousands. Yes, it utilizes plutonium, but an isotope of plutonium that isn’t used for weapons. But more importantly, that plutonium never leaves the site. Nothing leaves the site but the fairly short-lived waste. No need for a “yucca mountain”, no need for storage of spent fuel rods. We are being stupid with nuclear power for no good reason. Wind/solar isn’t the answer when a hail storm or hurricane takes out the entire generation capacity for a region. I have never heard of a nuclear or even a coal power plant being taken out by a hail storm or hurricane. Turkey Point power plant did sustain damage in Andrew with damage to a smokestack and water tank of one of its conventional coal generators. Such a storm would completely wipe out all solar and wind generating capacity. We can NOT strategically rely on wind and solar to provide reliable power in the face of bad weather. We CAN rely on coal and nuclear to do that. Meant “conventional nuclear power”, not “conventional power”. Willis, thank you for a very interesting article and to (most of) the contributors for extra information and links. @ Bill DiPuccio – 50 years to pay off? I’ve just had solar PV installed here. My calculations are for a return (conservatively) of 6.6% p.a. but that is after subsidies. Without them I estimated 1.1% p.a. @ Mr.Cannuckistan – solar PV giving 55% of peak in winter? I checked out my neighbours results and for 2 quarters of the year the amount generated was about 15% of the amount generated. Of course, this was in Adelaide (South Australia) where the sun doesn’t shine as brightly as in Canada sarc now off/. @ Philip Bradley, Perth – largest wind farm off-line because of mice eating cables – would that be the Albany scheme? When I was visiting it, I noted the high Capacity Factor claimed as likely [41%] and put it down to the ideal siting. Since then I’ve found out that it actually runs around 32%, or less if the mice are hungry. NOTE heroic refraining from gags about Nature. Our local Sunday paper has a report today on the up-take of solar PV here under the old scheme of 44c per MWh generated (gross amount, not what went back into the grid). The highest % of houses taking this option was on the coast SOUTH of Adelaide in the ‘retirement belt’ at around 39%. obviously the “sea-changers” had funds available. The 13 other suburbs listed are all “mortgage belt” homes. It appears that the common reason was to avoid future electricity price rises, apart of course from greedy bastards like myself. For comparison purposes local costs are 10-11 c (overnight rate) 20-23c (Peak rate 27c) per KWh in daytime, but with rises of 16% p.a. likely. Not least of the reasons for rising costs is the necessity to up-grade the local parts of the grid because all those solar cells start (& stop) generating at the same time, leading to higher voltages in the lines and at the local sub-stations. SA also has the highest % of wind capacity installed of any state in Aust. (approx. 20%) which was the boast of our recently departed (and “financially, scientifically, and truthfully challenged”) Premier. The backup costs to that don’t help, as the local CF is 20-25%. It would be interesting to see a comparison of retail electricity rates versus the percentage of “renewables” installed. My guess is that they both climb together. Re the difficulties of costing power, I point out that the overnight rate from coal stations is enough to keep the stations running under load ready for the peak demand to come (once the sun comes up). But as a guide to the real cost that figure is a guide. As is the French price to the UK for nuclear power in their last 4 winters when those thousands of wind turbines weren’t generating. That was 69 euros per MWh or roughly 50c (US) per KWh. Apologies for the lengthy response, but most of you will read this on Sunday, unless you’re outside maintaining your renewables. The majority of the land area of the US is subject to hurricanes, tornadoes, blizzards, and hail. Some places are subject to all four. Let me put it this way: if your turbine and solar panels get destroyed, you are going to want the grid power to be there. So … the grid power has to be there to back you up. If we place a significant portion of our power generation into modes that are subject to being wiped out by the weather, that is just plain stupid. “crosspatch: Let me put it this way: if your turbine and solar panels get destroyed, you are going to want the grid power!” And if your big power plants are destroyed by an earthquake (as recently in Japan) than you are more than happy to have some solar panels on your roof. BTW I do not believe that solar power will never by competitive for consumers, the prices are falling and with roof-top installations there are no major infrastructure costs. Furthermore as the EIA already pointed out in the past due to growing energy demand worldwide we need all sources, so it is not gas, coal or solar and wind but all of them together if we want to maintain our standard of living. kakatoa says: December 3, 2011 at 5:32 pm). ———————————————————————————————- Doesn’t look like a very good deal to me. Based on current prices for grid tied systems, your cost would be in the order of $40,000 (ignoring and federal and state cash backs you received.) Interest on that money at 5% per year is $2000. At 20 cents a kW, you don’t even earn enough to pay the interest on the capital. You may as well have invested it in a good dividend stock and used it to pay for the extra electricity. Even at your tier three four and five levels, it takes 23 years using a 5% discount rate and 30 cents a kWhr to get to the break even point, at which time you will be replacing panels and inverters and tie systems, if not before. And how is your system doing in the Santa Ana winds I wonder. Just curious as my panels sure don’t like dirt, rain, clouds or snow. Maybe turning off the air conditioner and opening a window would be the most cost effective. (I have a 3kW gen set on my horse trailer and it barely runs the air conditioning – 3 kW – that’s the equivalent of turning on my electric kettle, my two slice toaster and a couple of light bulbs, the radio and the range hood – anything more and the overload kicks the gen set off. Just so people get and idea of what we are talking about.) So good for you if you paid less or got cash backs from governments, but I can’t make the numbers work for me without subsidies of some sort. Someone is paying. Thanks, Willis, especially for raising the point of peak load vs. base load, and how solar and wind need backup gen capacity: that is so often forgotten. I’m a bit more optimistic than you on solar, though not much. It looks to me like the infrastructural costs (costs other than the solar cells) could change, depending in the infrastructure needed for various hypothetical systems. Certain types of mounts don’t cost much, and access roads can be dirt, etc, so I’m hopeful that in the future, the price might be competitive. However, we’re nowhere near there yet, and IMHO subsidies make it less, not more, likely that we’ll get there. I don’t mind subsidizing true research (a very small cost compared to subsidizing generation stations like we’re doing now) but I would like to see the subsidy of production end. Speaking of solar; one thing I hate are solar garden lights. Sure, free electricity, but they cost many times any savings in battery replacement every year or two, are unreliable, etc. They are great for spots where it’s too hard to run wiring, but nowhere else. I’d much rather take the time to run low-voltage wiring. (plus I get decent lighting that way; the solar ones are so often uselessly weak). h DirkH writes “That’s a straw man. Energy demand is flexible.” Only to a point and thats easy to say whilst there is enough energy around to meet demand. Every single (man made) thing you can see around you has considerable energy input in its manufacture its not just the number of kWh you use in your home every day. More importantly there simply is no viable alternative for oil at all yet. Yeah people all go on about possibilities like coal fracking and whatnot but the fact is that it takes significant time to get that kind of infrastructure up and running and I would suggest that politically it will be difficult to get moving at all let alone “in time”. how much electricity even the most diversified wind grid will actually produce at any instant – starting from zero. In that way the debate about wind fran yield factors is irrelevant. It makes no difference whether the yield is 25% or 30% or even whether it can be increased to 50%. it matters that whatever the average yield over periods of time, the probability distribution of output at any instant will be little changed, be too concentrated around zero and too unknown. Like the solar option, a game changing, but as yet unknown, technological advance is needed before it makes any economic sense to invest in wind and only subsidy and regulation will encourage it until that happens. I correct myself…When I referred to Fracking coal, I actually meant the Fischer–Tropsch process which essentially turns coal into oil. In theory It could extend our oil supply but will take a monumental effort and significant time to get production to the levels required globally. And an even larger effort politically to make happen. Even though I dont believe CO2 is the evil that the AGW people believe it is, I do believe we are going down the right path chosing renewables over continued reliance on fossil fuels. Especially distributed small scale solar which IMO is one of the best options forward. On the opening graph, nuclear (at 11c) is placed below biomass (12c) and advanced coal (12c). Something is wrong. When I look at the costings we used to review very frequently in the early 1980s, the dominant factor was realley the physics of energy density. This has not changed greatly. Solar panels might have got a few percent more efficient, but that’s a drop in the ocean as you note. The main deficiency I see in these figures is the huge burden that has been added to nuclear. It’s a synthetic blend that emerged step by step to placate green opposition. There are large social compliance costs, like the cost of site investigation, like pre-paid insurances with huge premiums, like mandatory charges to study waste disposal that have been collected but never utilised into a practical outcome, etc etc. I’d also note the Moore’s law comment about solar prices dropping and apply it to nuclear technology. The new generation plants should produce more cheaply than the older. They are simpler. Then you need to factor in the longer practical operating lives of nuclear plants, which are not estimates now, but reality. Taking all this into account, I’d tend to put modern nuclear TRUE costs about where hydro is. Historically, this is where it has been, if artificial social costs are removed, since the 1970s. Scientists have to find ways to solar power viable for everyday use. With the prices of petrol and gasoline at an all time high, not to mention the side effects of these on the environment, alternate fuels sources are the need of the hour. Rob Goodwin says: December 3, 2011 at 4:38 am Goodwin has cognitive dissonance. He is a solar power advisor. His outlook is short term (inal). From UK elswhere: On October 31st 2011 the government announced that they will be reducing the feed in tariff by over 50%. All existing enquiries had to be installed by December 11th to qualify for the feed in tariff at the current rate. This means that we are no longer able to take anymore enquiries for free solar. However a paid solar system is still a great investment even with the feed in tariff at the reduced rate. CRETINS If free solar does return you can still register your details and we will contact you if there are any changes. FAT CHANCE LOSERS! Please accept our apologies if you are here from any publications that have already been distributed. EIA is skewing coal numbers. Also, it is laughable to believe that CCS can be done for a penny. Every coal plant is going to deepwell CO2, or are we going to build pipelines crisscrossing the US to injection stations? I’m not aware of any real CCS other than heavily subsidized demonstration projects. Anyone who thinks it can be done for just a penny more (per kW) ought to be drug-tested. Deepwell CO2? I suspect that bs may be reconsidered. The other point to keep in mind is that the value of wind and solar energy is different. The value of wind energy is avoided fuel consumption at the balancing plants. Essentially the cost of gas. The value of solar energy is somewhat higher because the output is more predictable. You don’t have to have as much back-up power for solar plants, so the value is avoided fuel consumption plus some avoided capital costs for other plants. Also, it should be pointed out that the utility scale solar industry is still finding it’s feet. The biggest plant in the US is 48MW – tiny. So, they still build them basically as a collection of commercial sized plants, with 600VDC collector strings and 500kW inverters. Very little balance of plant optimization has been done. Shorter version: “free” energy is too expensive if the O&M and capital depreciation is too high. “Too cheap to meter” will never happen because most of the cost of a delivered KWH is in the delivery system. The only theoretical possibility of that ever happening is very localized generation, and there’s nothing fitting that description on the horizon. Now WHY were the Fukushima plants destroyed? The ONLY reason was because they were an obsolete design that required outside power to dump decay heat after shutdown. Of particular irony was that the explosion at Unit 1 probably resulted in the cascading failure of Unit 2 and Unit 3. Had unit 1 not exploded, chances are good that the other two wouldn’t have either because the explosion of Unit 1 cut the power cables that had just been laid to Units 2 and 3. Units 5 and 6, being of newer design and being placed higher up, suffered no catastrophic damage. In fact, their power generators were still operating (though not needed in those designs to dump decay heat but those units were not operating). The ultimate irony is that Unit 1 was slated for final shutdown for dismantling within three weeks of the quake. Had the quake waited two more months to occur, we wouldn’t be having this discussion at all. Unit 1 would have been in cold shutdown already and would not have exploded. Now the larger point is that if we installed MODERN plants that do NOT require external power and pumps to shed decay heat, we would be immune to the sort of problem that knocked out Fukushima Dai-ichi. It was not the quake but the tsunami that took out that installation by taking out the generators required to run pumps and operate valves. Modern plants do not require generators to run pumps or operate valves for emergency cooling. Fukushima Dai-Ni, on the other side of town, experienced the same quake but those units were of newer design that also did not require external power to run pumps. Those units had steam turbine pumps that could use its own decay heat to power the pumps. Modern plants don’t even have pumps! They operate on convection, gravity, condensation and evaporation and can remove decay heat passively. We need to replace the old plants with new ones and build more of them. For emergency cooling that is, the emergency cooling is completely passive. And one other detail – if solar is cheap enough, a small penetration might be feasible in certain markets because the output coincides with the daily peak, and it’s output is valued at peak prices, which are much higher than average. Not only do you get a premium price for the power, you don’t need storage. Once you get past a certain point though (probably 5-10%), the economics go downhill rapidly. Manoj says: December 4, 2011 at 4:02 am . . . alternate fuels sources are the need of the hour.” Maybe the next hour, or the next, or the one after that. This morning it is a bit chilly where I live. I have an all electric house powered by falling water on the nearby Columbia River. That is true of my neighbors, who also burn wood as an alternate fuel. The only thing likely to change in the next ten years is who cuts and carries the wood. Seriously, study the possibilities for massive alternative energy. What technology can be applied? How fast can it be built? Think materials, workers, land, transmission, among other factors. Would anyone object and file a lawsuit for a project? Even, say, a green wind project? There is historical precedent for a transition from one form of power to another. See: “The Centrality of the Horse to the Nineteenth-Century American City,”* an article by Joel Tarr and Clay McShane explaining the serious environmental hazards horses presented when used in large numbers and how that related to the emergence of the automobile. Still, these things take time. Put in a different light: “environmentalists” are making it much MORE likely we will experience a disaster like that in Japan by inhibiting the replacement of old plants of 1960’s/1970’s design with new plants of more modern design. A modern plant could weather an earthquake such as we experience in California (a “strike/slip” quake of up to about M8). By delaying the construction of modern plants, they are causing older plants subject to the same sort of failure as the Fukushima plants to operate longer. All nuclear plants are not the same. Modern plants such as the GE ESBWR and the Westinghouse AP1000 are not only safer, the are simpler to build and maintain. Most of the cost in building and operating nuclear plants is due to external costs imposed by regulations that are not directly related to how much it costs to actually generate power. One example is spent fuel. We are not allowed to reprocess it into new fuel (typically a fuel rod is “spent” when only 5% of the energy potential of the fuel is expended leaving 95% that could be reprocessed and used). Not only are we not allowed to reprocess them, we aren’t allowed to dispose of them either so we end up with massive amounts of spent fuel that must be guarded. It is stupid! Sometimes I believe that we have become a nation of emotionally driven idiots. Germany’s ahead for some more trouble w/ renewables. According to the Handelsblatt (in German) Poland is about to repel german renewable’s export in order to protect their power plants from ramping down – forcing us germans to eat our own dog food. Of course they won’t do that. They want compensation. I can’t blame them. First of all the nuclear cost figure is pure fantasy. It might be that low after you force the captive rate payers to eat the cost overruns! The solar pv cost figure is also fantasy because we still let analysts get away with inflated industry averages that are falling fast and includes the costs from wannabe startups with Federal loan garauntees and no future. In short, misguided statistics equates to more lost time on rational energy policy choice.. I was surprised by the entry for wind at 10 cents. That is close to competitive but first we need to solve a few problems like storage of massive amounts of DC electricity cheaply. Without that no intermittent source is practical on a large scale. If you can get rid of the requirement for base load energy as backup you can make a much better case for wind/solar etc. If you can’t? Well you are sunk. It is very similar to your house. Solar and for that matter wind, micro hydro etc really only make sense when you go off grid completely. The reason for that is all the other charges on your bill. In my case a full 2/3rds of my gas, water & electric is not consumption based. To pay back solar would take forever unless I would disconnect and save that 66% of my bill. That costs a lot more up front because now you are looking at serious amounts of storage for electricity. I’m also amazed at how much emphasis is placed on big projects. Yes there are advantages to scaling but like Willis has pointed out there are a lot of other costs. So when they build roof tiles with solar built in that will be a major advantage because you are going to get the roof tiles installed anyway. would love to see some serious investigation about that approach. TimTheToolMan says: December 4, 2011 at 1:22 am Didn’t I just go through and show that “distributed small scale solar” is about three times the price of grid power? Didn’t I just show solar is very unlikely to ever be less than about 12¢ per kWh no matter how cheap the panels get? Didn’t we discuss the fact that these numbers don’t include the necessity for spinning backup, with associated capital, fuel, and operations costs that aren’t included in the 12¢ per kWh? And that as a result solar will likely never cost less than 15¢ per kWh or so? How on earth can you twist that into thinking that solar is “one of the best options forward”??? Grid-connected solar is a lousy option forwards, hugely expensive and a bad choice. w. ChE says (emphasis mine) December 4, 2011 at 9:05 am Thanks, ChE. There’s two parts to transmission costs: connection to the grid, and moving (“wheeling”) the power around the grid. The “wheeling costs” (the costs to move a kWh around the grid) are a small part of the cost of a delivered kWh. They around a penny or so per kWe depending on the grid. (A “penny” is the US 1¢ coin.) Getting power to the grid is not all that expensive either. The most expensive connection to the grid is for offshore wind. The EIA puts that cost at 0.6¢ per kWh (six tenths of a cent). So your statement in bold above isn’t true. Transmission costs are typically a small part of the cost of a delivered kWh. w. ChE says: December 4, 2011 at 9:11 am First, what part of “solar is not cheap enough, and won’t be cheap enough any time in the near future” is escaping you here? Second, there is an illusion, likely fostered by the power companies, that peaking power is hugely expensive. In fact, it is only as expensive as the most expensive power in the mix. Typically, peaking power is provided by gas turbines. Looking at Figure 1, these come in at 11¢ per kWh. The other problem with “replacing” peaking turbines with solar is … what happens when a cloud passes over your fifty megawatts of solar? You need to have extra spinning backup power for solar. So you need even more gas turbines to prop up your solar megawatts. Finally, huge gas turbines don’t like being speeded up and slowed down, up and down every time a cloud passes over your green power plant. The thermal shock greatly shortens their lifetime, which again adds to the cost of solar. Summary: 1. Solar is far too expensive to replace even peaking power. 2. Solar is unlikely to be cheap enough to replace peaking power any time in the near future. 3. Even if it were cheap enough, the up/down cycling and the need for spinning backup make it useless for peaking power. In short, Solar electricity = brilliant plan Grid-connected solar electricity = very expensive bad plan w. I spent most of summer in Denmark and Germany and hardly saw the sun. The best place for solar panels in these regions is inside the barn so as they don’t get dirty and you can shine lights on the panels to generate electricity. Australia is probably one of the best countries for solar panels but the unit cost for a kWH would need to more than triple to make solar panels and wind turbines viable. Our federal government here in Oz us tell us in one of their brochures that once sufficient solar panels and wind farms are up and running coal fired power stations can be decommissioned.The availability rating for electricity generated from coal fired power stations is currently around 99.95%. I dread to think what that rating would fall to if we disconnected all coal fired generators. crosspatch says: December 3, 2011 at 8:10 pm Gail, wind does not work. Well, not on a large scale….. _________________ It has always been a scam. I looked into it a while ago. Wind might work small scale if you live on a windy ridge (I do) and can use it to pump water between two ponds with a hydro-generator between them. Otherwise they only make sense for pumping water into livestock tanks in remote locations or grinding grain in third world countries. The big problem with all of it is the town zoning laws. A large corporation can do what a home owner can not. Which is probably lucky or we would have small scale bird shredders littering the landscape. TRM says: December 4, 2011 at 11:46 am … also look into this one too. These people seem to have done the engineering: I talked to a guy yesterday who sells small heaters. He knows of two houses in my area with the buried pipe heating/cooling system. Now if I could only convince my Community College to use my house as a demo project. Windmill with a two pond hydro power unit for self contained electric and the geo-thermal for the heating/cooling maybe I could get a grant…… (snicker) Willis, The folks in Marin are a bit unhappy with how PG&E is allocating it’s cost to provide service- .” as noted here- : “PG&E charges ahead -Is utility manipulating rates to undercut Marin Clean Energy?” They are likely upset about how much PG&E is allocating to transmission and distribution- “Charges for your electric service include both generation (electricity) and transmission & distribution (delivery of electricity) components. PG&E’s transmission & distribution rates are applicable to all customers. PG&E’s generation rate is not applicable to customers who do not receive their generation from PG&E, such as CCA customers or Direct Access customers. Generation rates for these customers are determined by their energy service provider, although PG&E may also include PCIA and FFS charges to reflect the full cost of your generation services. See definitions below and on Page 2 of your bill.” . This quote is from The non-generation part of the A-1 (business) rate is 65% of the costs of a delivered kwh. This percentage is similar the allocated costs, per my yearly true up bill from PG&E, which was 36% for generation for 2009 and 2010- my allocation for generation droped to 25% last year. My allocation for public purpose programs jumped last year to 24% of my yearly bill (vs 16% the year before). The unbundeled rates for residential (e-1 rate schedule) is delineated here- It’s been many a year since my cost accounting days, but the breakdown in the costs noted in the schedule don’t seen to have any relation to the actual cost to deliver electrons to end users. Your thoughts on the allocations would be more then welcome……….. TRM says: December 4, 2011 at 11:46 am .” Electrical generation is not a great concern for the United States. Transportation fuel is the problem. We could end the recession in the U.S. by easing the price burdens on energy production which is exactly what Rick Perry intended to work towards if elected president. “I was surprised by the entry for wind at 10 cents.” You would be if you believed most of the authors on this website who have a knee-jerk ideological opposition to anything and everything that is favored by the evil environmentalists. Like most things in life moderation is the key. Wind energy is no panacea but carefully managed in suitable locations it can comprise perhaps 10% of electrical generation at competitive prices. “That is close to competitive but first we need to solve a few problems like storage of massive amounts of DC electricity cheaply.” Either that or get better at predicting demand and production. That’s essentially saying we need better weather predictions. Better wind predictions so we know how much wind power is available at any given time and better temperature predictions so we know how much heating/cooling demand there will be. Given adequate information you can schedule adequate spinning reserves in a cost effective manner. “Without that no intermittent source is practical on a large scale.” Certainly not as a sole source but I’d have to say with over 10 gigawatts of nameplate capacity wind energy is already large scale in Texas and still growing. ” If you can get rid of the requirement for base load energy as backup you can make a much better case for wind/solar etc.” A better case, yes. “If you can’t? Well you are sunk.” No you aren’t “sunk”. You have more constraints. “It is very similar to your house. Solar and for that matter wind, micro hydro etc really only make sense when you go off grid completely.” I found the complete opposite. Batteries are over half of the total cost of an independent system! Solar PV would be cost-effective for me (owner-install) at half the current price of grid-tie electronics and solar panels. It becomes far more difficult if I’ve got to maintain a huge battery bank to supply juice during nights and cloudy days. Prohibitively so. With a grid tie I can, on most days, do all my heating and cooling while the sun is shining and any nightime or cloudy day usage I can draw from the grid. Simply heat or cool thermal mass during the day and use that to keep the temperature constant during the night. I already leverage thermal mass to a large extent using earth-berm but I have an advantage in that the year-round temperature a meter or more underground in my location is 72F which is perfect. I excavated a chunk or north facing hillside with only one long wall and the roof not backed by earth. Very little energy is needed for heating or cooling. The inside temperature absent any heating or cooling changes very little from one day to the next and the closer the average daily temperature is to year-round temperature the less change there is inside. “The reason for that is all the other charges on your bill. In my case a full 2/3rds of my gas, water & electric is not consumption based.” I have a water well, septic system, and propane tank at my primary residence so there are no non-consumption charges in those. Non-consumption electric charge is only 20% of my bill on average not 66% as in your case. You either use very little electricity or your electric company is taking you to the cleaners for service availability charge. Mine’s bad but it’s still only $22/mo. “To pay back solar would take forever unless I would disconnect and save that 66% of my bill. That costs a lot more up front because now you are looking at serious amounts of storage for electricity.” If I could generate electricity at half the cost of current photovoltaic cost/efficiency and I could sell it at retail price I could make a living at it. Unfortunately my electric company sells to me at retail price and buys back from me at wholesale price if I generate more than I consume. Given that every kWh I don’t purchase from the grid I save full retail price then with modest cost/improvement in PV panels it would be worthwhile to at least produce as much as I consume. “So when they build roof tiles with solar built in that will be a major advantage because you are going to get the roof tiles installed anyway.” Solid state electronics have a history of rapid price declines. PV seems to be a bit of an exception and grid-tie electronics are ridiculous in price. Economy of scale can probably drop current prices in half. .” That’s essentially what I do but it only works in locations where average year-round temperature is in the same ballpark as room temperature (72F). Where I grew up it’s 52F which makes it much less effective. David L. Hagen says: December 3, 2011 at 7:16 pm I’m sure Willis has “moved beyond” this which essentially means he’s convinced he’s right and contrary evidence is therefore not worth his time to consider. Willis practices dogma not science. That said, even if solar can be competitive with natural gas, that doesn’t really solve any problems unless one considers CO2 from fossil fuel combustion to be a problem. A far as I can determine CO2 is a benefit not a problem. We need CHEAPER sources of energy to make progress. A same-price replacement is worth nothing AFAIC. You can run by a design for a Liquid Fluoride Thorium Reactor (LFTR) and know it won’t be half the price of a Gen III Pressurised Water Reactor(PWR), which are the ones being planned for and built at the moment – that takes it to, say, 5cents. Also the thorium fuel costs, operation and maintenace will be half that of PWRs – 2 cents. All in all LFTRs will be as cheap or cheaper than CCGTs (and they can load follow as well). See the heading to this blog to get an instant perspective on the environmental effects of obtaining our energy from: Coal – Uranium – Thorium: Yes; God forbid you should look to ‘dropouts’ like Wozniac, Jobs, Gates, Dell or a Henry Ford to find reasonable market-workable solutions (read that as: economically viable and producible). ALWAYS look to your ivory-towered, pointy-headed academics to ‘solve’ your problems (ISN’T that what got us into this mass in the first place?) /Not even sarc . RE: Dave Springer: (December 4, 2011 at 1:43 pm) “We need CHEAPER sources of energy to make progress. A same-price replacement is worth nothing AFAIC.” On the assumption that economically recoverable geo-carbon energy is being exhausted, a same-price replacement is better than a high-cost replacement. Bravo, Willis! You’ve captured the essence of why solar is a hopeless cause. The bottom line is a pathetic energy density that precludes the technology’s ever achieving “economy of scale”. There is simply too much physical material required to produce too little power at a horrible capacity factor. Wind power is afflicted, to a less extreme degree, with the same problem. I’ve only a three of things to add. First, I think the levelized figures for solar are, in reality, considerably higher than those shown in the chart. Second, the levelized costs of CCGT (natural gas turbines) are dominated by projected escalation of future natural gas prices. I’m betting those prices will be less that projected in the U.S. because we are bursting at the seems with newly proven natural gas reserves. At today’s natural gas price, CCGT pencils out at under 4-cents. Third, wind industry experience in the U.S. indicates that O&M costs are much higher than those shown in the chart. So many moving parts for so little power is eating their lunch in maintenance expenses. Willis writes “Didn’t I just go through and show that “distributed small scale solar” is about three times the price of grid power? Didn’t I just show solar is very unlikely to ever be less than about 12¢ per kWh no matter how cheap the panels get?” Yes, Willis, you did say those things. Based on an analysis in today’s environment. But IMO you’re showing a surprising lack of vision on this matter. Nothing comes close to fossil fuels for energy density (except nuclear and location specific hydro and geothermal perhaps) so what are you saying we should do? Stick with fossil fuels because they’re the most economical? Here is a recent video of a talk given by Dr. David LeBlanc on the design of thorium-based nuclear reactors with low potential energy cost. As far as I know, thorium-nuclear is the only energy resource that can meet our current needs indefinitely. David LeBlanc – Potential of Thorium Fueled Molten Salt Reactors @ TEAC3 “Dr. David LeBlanc explores the diversity of Thorium Fueled Molten Salt Reactor design options, and their rational and value. “Presented at the 3rd Thorium Energy Alliance Conference, in Washington DC.” Uploaded by gordonmcdowell on Nov 27, 2011 12 likes, 0 dislikes; 222 views; 20:13 min. daryanenergyblog A critical analysis of current and proposed future nuclear reactors designs Part 8 – The Molten Salt Reactor concept Dave Springer says: December 4, 2011 at 1:43 pm What is it with you and nasty, unprincipled, untrue allegations, Dave? Less than 24 hours from David Hagen’s post to yours where you accuse me of “moving on” … you are a jerkwad, do you realize that? Why are you always so unpleasant? Do you think acting like that gets folks on your side, that it leads them to want to listen to you? I did not answer David’s post because I didn’t see the relevance of it to the subject under discussion. I mean, it’s great that the DoE wants to fund R&D into the costs to cut them by a factor of three. Get back to me when it’s halfway there and we can talk about it again. Until then it’s just another government pipe dream, and I’ve seen dozens of them come and go. So why should I spend time discussing this latest one? And what difference does it make to you? Have you appointed yourself the arbiter of what is worth answering? There’s 233 comments on this post. I pick which ones I answer. You don’t like my choice? Fine. Free country. You want to bitch and piss and moan about my choice, and accuse me of assorted eco-crimes up to and including mopery on the skyways? Typical. Get a life. w. TimTheToolMan says: December 4, 2011 at 6:10 pm Yes, of course we should stick to fossil, until we find something more economical. Are you saying we should switch to other fuels because the other fuels are uneconomical? w. Willis writes “Yes, of course we should stick to fossil, until we find something more economical. Are you saying we should switch to other fuels because the other fuels are uneconomical?”. Re:TimTheToolMan says: December 4, 2011 at 11:01 pm .” If you simply let market economics decide when a better alternative to fossil fuels is “ripe”, you avoid all that unnecessary pain in your wallet at tax time and unnecessary pain in your butt that premature adoption of the bicycle commute will engender. I’m in favor of a “first adopter” mentality, so long as I’m not forced into paying the freight for your farsightedness. Claude writes “If you simply let market economics decide when a better alternative to fossil fuels is “ripe”,” Market forces will be driving it all along the way, but market forces alone mean the path of least resistance and that isn’t necessarily the best for long term goals. There is certainly a role for subsidy and regulation to get things done that are in the best interests of everyone and not just those doing the driving. Take fishing as an example, if it weren’t for regulation and society self imposed responsibility, we’d probably have essentially fished out the oceans by now. TimTheToolMan says: December 4, 2011 at 11:01 pm We have made every previous energy transition without government intervention or direction. We are currently transitioning away from coal and towards natural gas without direction. Now, you want to direct us to renewables for the grid. You’ll excuse me if I don’t want to pay for your good ideas. Truly not interested in shelling out for some green ideal. Don’t mistake me, I’m a man who loves renewables. I wrote the Peace Corps manual on windmill construction. I’ve lived off the grid on solar. Renewables rock … just not for the grid. Energy is development. Taxing it or making it more expensive hurts the economy, and in particular it hurts the poor. A rich man doesn’t care how much he has to pay for electricity or to fill his gas tank. Price hikes on electricity, the kind you are blithely proposing in the name of some holy goal, hurt the poor. You may not care about hurting the economy and or about hurting the poor. I do. w. Willis writes “We are currently transitioning away from coal and towards natural gas without direction.” And thats a good move as far as I’m concerned, but not a sustainable one. So all the effort and costs we pay for transitioning towards natural gas will ultimately need to be made again when we transition away from it. And meanwhile, looking at your picture, the price is mostly related to the gas itself and thats only going to increase. I like PV solar because the more of it that is out there, the lower the overall running cost. The cost is born up front and whilst you may not immediately see the benefit of that, I certainly can. Society as a whole doesn’t share my point of view because we’re living very much in a “I want it now” society. TimTheToolMan says: December 5, 2011 at 1:10 am Thanks for the reply, Tim. Your solution to possible future transition costs is to triple the price of electricity and maintain that indefinitely? Yes, but at present the supply is increasing. It will be around for decades. Certainly some time in the future we’ll have to replace aging gas plants with something. By that time it may be an energy source undreamed of at present. But tripling current energy prices to solve that possible future transition cost? Sorry, that’s spending a dollar to save a dime. You have no evidence that the savings will be significant. Solar is not some new technology with huge savings to be realized. Much of the low-hanging fruit was plucked a while ago. There will be incremental savings, but not huge savings. And you need huge savings to bring it into line with other power sources. No, the cost is not “borne up front”. It is generally taken as a long term loan, and paid back by increased electricity rates over the long term. It is a continuous and unnecessary burden. You have not replied to the issue that any tax or increase of energy prices is a hugely regressive tax hitting the poor the hardest … Thanks, w. Re:TimTheToolMan: (December 4, 2011 at 11:01 pm) “One way or another we’re going to be transitioning away from oil and I’d prefer the significant related expenses and efforts were directed towards renewables rather than increasingly ramping up our mining and processing of coal.” I will agree that we should be looking to find a replacement energy resource as the limited stores of economically recoverable geo-carbon energy are depleted. But traditional ‘renewable’ energy resources, those ultimately based on energy from the sun or geothermal energy, I think must be ruled out as they never have been able to support more than a small fraction of our energy needs. One might ask; how large would a solar energy farm have to be if it were to supply the total energy now used in the state of California? How many people would be required to keep the cells clean and functional? What would we do at night or in cloudy weather? How much expensive copper would be required to link all those cells together? I suspect that all we can look forward to is a reversion of population and lifestyle back to the 1880’s, if we are going to be limited to energy from the sun as our primary energy resource. In that case, government officials who see this coming might be forced to put in place various unpleasant policies to facilitate an orderly population reduction. The only energy resource that I see on the horizon that has any real likelihood of replacing ‘Carbon Power’ at our current rate of use is energy from thorium. ”Renewables alone, are not going to power this economy,” Mitt Romney. ferd berple says: December 3, 2011 at 7:52 am “However, this won’t happen, because only the rich can afford the $50,000 investment, which means the poor in Ontario are paying the rich to install solar power.” As it is allways, with subsidies. A few gets a lot from the many. A funny everyday story from Norway: A few years back a new independent company started producing milk much cheaper than the govmint milk. The govmint forced them to put on a tax on it, so the price became equal to the govmint milk. Oh yes. Willis writes “Your solution to possible future transition costs is to triple the price of electricity and maintain that indefinitely?” No, because I dont believe the cost is actually that high in the longer term. These studies are always based around what is known today and one can only imagine what might happen in the future as a result. But inevitably they dont form part of any projection as those changes tend to be speculative. I’m speculating. A large component of the cost associated in those figures is the cost of energy itself and so drops in running cost are of a long term benefit in keeping costs low even if they initially cost more. I dont think its always easy to see that. So for example for argument’s sake imagine if we already had 100% solar PV then what is the cost to create another panel? Much lower than projected as there is a very low energy cost component and thats an obvious example. Its not always easy to see the related lower costs associated with all the activities surrounding the activity as well as the activity itself. Willis then goes on “You have no evidence that the savings will be significant.” You’re right. However I do believe the world where energy is not a resource based commodity (beyond initial manufacture obviously) would be a better one for many reasons eventually both economic and political. Thats all ideal…I am more a realist than that, however, and fully expect we will go down the path of least resistance because thats what we always do. I can still have my say on the matter though. @Willis “But with hydro (or almost any other conventional technology) you only need to maintain one really big generator on the ground.” Really? [sigh] [SNIP: Policy -REP] [SNIP: Policy -REP] (attempt 3 to get this comment posted) [REPLY: It will NOT posted. You have a grievance. Click on the ABOUT tab under the WUWT graphic and then click on “contact”, but this propensity for flame wars stops. -REP] Spector says: December 4, 2011 at 6:24 pm “Here is a recent video of a talk given by Dr. David LeBlanc on the design of thorium-based nuclear reactors with low potential energy cost. As far as I know, thorium-nuclear is the only energy resource that can meet our current needs indefinitely.” The we’re screwed. .” Unfortunately it IS cargo cult science. It’s far too difficult to process solid fuel thorium. That leaves liquid fuels which in general is liquid flourine salt a.k.a. “LFTR” designs.. This is simply a bunch of people trying to make a fast buck by getting research funding and ignorant idealists who are willing to give it to them. But hey, it’s a small step ahead of fusion (cold or hot) in practicality but it’s a long way behind solar (hydro, wind, biomass, biosynthetic).. There is no basic discovery needed to accomplish this which means it’s an engineering problem not a science problem. When you have a situation like the LFTR where there’s no known material that can meet the design requirements it means there’s basic discovery involved and discovery of a novel material that doesn’t exist cannot be predicted nor even guaranteed. Sythetic biology on the other can be guaranteed because the technology and materials required already exist in nature and just need to be recombined rather than be invented. The recombination technology is in its infancy but is beyond proof-of-concept. The first completely artificial genome that brought a lifeless bacterial shell devoid of DNA back to life upon insertion happened a couple of years ago. It’s only a matter of time until the painstaking, error-prone process of creating synthetic organisms gets cheap enough and fast enough so that the trial-and-error process of recombining various desireable functions of different organisms into one super-efficient hydrocarbon fuel producer will meet with success. The Venter Institute is leading the way. Any thorium reactor, even if it could become economical someday, is almost certain to be obsolete and uncompetitive by the time it could be brought online commercially. Even the most optimistic estimates put a working, commissioned, commercial thorium reactor at least 20 years in the future and then it has to operate for another 20 years to recoup the cost of building it. That’s 40 years altogether and if something much cheaper comes along in the meantime it means anyone who invested in thorium will lose money and investors, at least the smart ones, don’t tie up money for that long without a really good chance of seeing substantial profit from it. Therefore thorium reactors are something only governments will possibly undertake. The United States government already built one and the classified details of its operation over ten years back in the 1950’s and 1960’s isn’t inspiring any new investment in it. One might wonder why the only nation in the world with actual experience with LFTRs is panning the notion of taking a second look at it. Willis: According to the Globe and Mail — you’re wrong….” Now when have they ever been wrong…??? However, Mr. Robertson noted that Ontario’s FIT program – like similar schemes around the world – was designed so that the price developers get for the renewable power they generate falls as the cost of producing it declines. And of course I though they were signing 20 year contracts — silly me… Now, I’m off to see the Wizard… for reliable information. ;-) Colin Megson says: December 4, 2011 at 1:59 pm Too bad there’s no known material that can simultaneously resist the corrosive action of 700C molten salt and high neutron flux. Other than not being able to build pumps and plumbing to shuttle the liquid fuel around they’re a really cool item, huh? Sort of like electric cars are really great if only there was an affordable battery with the power/density of gasoline. What part of “there is no known material that meets the critical design criteria for pumps and plumbing” do you not understand? coldlynx says: December 3, 2011 at 2:00 am “Never thought I would disagree with You Willis, but on this I do.” You probably hadn’t realized that once Willis reaches a conclusion then to him it becomes dogma. .” Yes of course. A minimal amount of research shows that the mass market potential is in residential and small commercial grid-ties. Cutting out the need to store power in batteries or some other scheme and instead selling excess generation back onto the grid with net metering is the way to go. Batteries easily double the levelized cost of the system and are only economically viable when the grid is so far away you can’t afford the cost of getting a connection to it. The decentralized nature of this also means that the current grid can handle a lot more capacity because when you have excess generation it will likely be consumed by your closest neighbors who don’t generate their own juice so it doesn’t add to the amperage on high tension long distance grid elements. .” Funny how it works that way with solid state electronics. One might have thought Willis was an astute enough observer old enough to appreciate what happened with radios, telephones, televisions, microwave ovens, and other solid state electronics. PV panels and grid-ties are no different except they have yet to see the benefit of economy of scale and adoption of industry standards. .” I’ve been in the computer business since the 1970’s but it doesn’t seem like it should take a rocket scientist to appreciate the price/performance curve in electronics from infancy to common household item. “And of the joy to get independent. A small scale revolution. ;-)” I’m not really interested in flipping off my electric company. It’s a cooperative to begin with and isn’t particularly offensive in any of its practices and serves an awful lot of rural customers with prices similar to densely populated areas where transmission costs are much lower. The retail price they sell power to me at is half the price they’ll buy it from me at which is reasonable. If PV generation price goes the way of other solid state electronics I’ll be able to sell electricity to the local coop at a profit. Now THAT would be cool. Re: Subsidies The latest (2010) EIA report is here. Renewables get 55% of all subsidies, most of it going to wind. About half of nuclears subsidies is in the form of R&D. RE: Dave Springer: (December 5, 2011 at 7:28 am) .” As far as I know, this was not a problem with the Oak Ridge demonstration unit. I understand the fluoride salts used combine fluorine with base elements to which it is more strongly attracted than almost anything else. Highly connected scientists from China are reported to be in possession of all data extant from the original Oak Ridge demonstration project and working on developing their own version. ”This is simply a bunch of people trying to make a fast buck by getting research funding and ignorant idealists who are willing to give it to them.” That is always a possibility. .” This sounds just like another form of solar power, perhaps ‘bio-solar.’ One might wonder how many solar petroleum trees or solar algae vats would be required to generate 100% of the power used by the state of California, how large an installation would be required, and how many people would be needed for their maintenance. My guess is that one would be lucky to collect, on average, 100 watts per square yard by such methods. With abnormal microorganisms used on a large scale, there is always the risk of them escaping into the wild and causing a massive change in the chemistry of the planet. “ The United States government already built one and the classified details of its operation over ten years back in the 1950′s and 1960′s isn’t inspiring any new investment in it. One might wonder why the only nation in the world with actual experience with LFTRs is panning the notion of taking a second look at it.” Of course, China is reported to have the surviving data from those experiments now. With multiple technical options available, there is often a tendency to standardize on the first method developed and suppress the development of all incompatible alternatives. I suspect the current administration looks at nuclear power as did Ralph Nader when he called it ‘poison power.’ Dave Springer says: December 5, 2011 at 6:24 am Dave, you can be as deliberately dense as you want, and pretend not to understand. But underneath, you know very well that my point is that one fossil fuel generator = a dozen windmills or more. So the maintenance is harder for windmills. Which you knew as well. Give it up, my friend. You are just making people point and laugh at you, and I doubt that is your intention. w. TimTheToolMan says: December 5, 2011 at 5:07 am Tim, your fantasies about future solar costs immaterial. Unless you can show that the cost is not that high in the long term, why do you believe it? And if you can’t show it … In any case, the cost right now for grid-connected solar electricity is about three times the cost of fossil electricity. For your fantasy to come true, every single cost involved in solar, from purchasing the land and installing the racks and purchasing the panels and every other part of the cost, would have to drop by a factor of three. I don’t see that happening, regardless of the strength of your belief. The industry is far too mature for that to happen. If you think initial costs are somehow immaterial and running costs are all that count, remind me not to let you invest my money … that’s just wrong. That’s exactly why we are discussing levelized costs, Tim, because they avoid all those problems. Please try to follow the discussion. Levelized costs use the net present value of future activities precisely to take care of the issue you are discussing. In addition, you can’t just add another panel to an existing system. To add that panel, you will most likely have to upsize the inverter, and increase the ampacity of the wiring, and purchase new, larger breakers, and increase the number of racks, and buy new land to put the racks on … you are a “toolman” and you don’t know this? That’s like saying “Once a house is built, it’s cheap to add a toilet, because a toilet doesn’t cost much”. Well, no, it’s not cheap to add a toilet, you may well need to extend the bathroom and to upsize the water pipes and to dig a bigger septic system and … That’s true. It would be great if energy were “not resource based”. It would be wonderful to inhabit a world where sunshine could be turned to electricity for free. And by the same token, a world where everyone was always nice and kind to each other would, as you say, “be a better one for many reasons eventually both economic and political.” Unfortunately, both of those are just fantasies. As you seem to recognize at the end … You might profitably consider that sometimes, the path of least resistance is the path we actually should be following … Thanks as always for your thoughts, w. I do wish some of you people would learn how to run a “life of project” financial analysis on a “levelized”, “discounted cash flow” or any other basis recognized as valid by investors and banks the world over. You would then see that capital costs, fuel costs, other O&M expenses and the time-value of money are all essential components of such an analysis. Any one of those elements can swamp all the others and render a project uneconomic. If you will take the time and make the effort to do that, you will recognize the foolishness of many of the pro-solar statements made in many of your comments, including the following grand prize winner: “The cost is born up front and whilst you may not immediately see the benefit of that, I certainly can.””. That’s the bottom line for unsubsidized central solar and that was the basis for Willis’ conclusion. It wasn’t an arbitrary or uninformed opinion. It was a rational conclusion that no amount of “hand-waving and posturing” should be able to change. Dave Springer says: December 5, 2011 at 9:30 am Dave, that’s a slanderous lie. I am one of the few climate bloggers willing to admit my mistakes for all to see, and I have done so quite publicly when I have been wrong. In other words, your petty jealousy has once again wrested control of your mouth away from your brain, and it is busy parading your childish vindictiveness up and down the town square. You really should do something about that, it’s not helping your reputation at all. You seem overall like a pretty intelligent guy … but man, sometimes you couldn’t prove it by your actions. What is your beef with me, Dave? What did I ever do to you to start you on this path of unending enmity? It’s not the subject matter, because no matter what I write, you always show up to do your gorilla trick. That’s the one where you defecate in your hand, and you fling it at me and the other guests … which is quite impressive in its own way, but likely not in the way you imagine, because at the end of the day, I’m not the one whose hand smells bad. w. Claude Harvey says: December 5, 2011 at 12:06 pm Thank you, Claude. Having worked as the Chief Financial Officer for a $40 million dollar a year company and run many, many cost analyses, I’d like to say that the amount of wishful thinking and fiscal ignorance shown by some of the solar advocates in this thread is shocking. Unfortunately, that kind of financial foolishness and misunderstanding is too common to shock me any more … w. Dave Springer says: December 5, 2011 at 7:28 am Dave, that’s an interesting idea. Do you have a citation for that claim? If the liquid fluorine compound is too corrosive to handle or pump or pipe, I find it difficult to believe that you are the first guy to notice that the corrosion problem was a deal breaker … or that scientists and promoters of the technology are ignoring that. I see references to corrosion in the literature, and people talking about how to control corrosion in LFTR reactors, and a special metal was developed to reduce it (Hastelloy-N nickel), but I don’t see anyone who claims it is a deal breaker. w. Willis, and all commenters Interesting analysis, I agree with your critiques of the information as published by EIA. There’s something else, though (I scrolled to Reply when I got ~1/4 way through comments, so forgive me if somebody else covers this) you hinted at only briefly, when you made the statement, “…will not be economically viable any time soon.” The numbers you published are static numbers, a snapshot taken at a particular point in time. For solar and wind to become competitive vs. fossil fuels, it follows we must see a rise in the price of fossil fuels? But what does that do to the production cost of not only solar panels but all the appurtenances you mentioned, as well as up-keep costs? My bet is all those costs increase, as well. I went to college and got a mechanical engineering degree with the vision of designing the replacement for fossil fuels. When I took elective engineering courses, I concentrated on solar thermal design. Our analysis found that the best you could hope for, with a cheaply built system, was a 10 year simple pay back. But, we logically reasoned, just as is reasoned today, that mass production will reduce the price of capital in a solar system, while the price of petroleum products will do nothing but increase. Well, apparently there was an invalid assumption in there someplace, or we failed to acknowledge the interconnectedness of prices of the choices, because here it is 35 years later, and still when I evaluate solar to replace fossil fuel generated electricity, I still get a simple payback in excess of 10 years, and the number I get today is even bigger than the one I got 35 years ago. You can change your “…anytime soon.” to NEVER. @Crosspatch (and Hultquist) Google frack acquifer mike g says: December 5, 2011 at 3:39 pm I find it laughable when people wave their hands at some huge mass of materials and say the answer is in there. Particularly when they write three words and mis-spell one of the three. Mike, if you have a point, and some citations to back it up, how about you just make it? Lay out your claim. Then bring on your best pieces of support, your best citations for your claim. Because I hate to tell you, but I’m not googling a dang thing. Google frack “acquifer” yourself. Fracking has been used for decades. As far as I know there are no documented cases of fracking disturbing the groundwater, but it’s a big planet. Out of the tens of thousands of times that fracking has been used, I suppose it is theoretically possible. You need to be cautious in your claims. There certainly are cases of natural gas in the groundwater. This is a natural and not unusual occurrence all over the planet, and as is always true in natural systems, is much more common in some parts of the world than other parts. So don’t bother sending us links to people lighting what’s coming out of their faucets. That happens not infrequently, and without any fracking nearby. w. This seems strangely appropriate. At 0:30 in this video, what do you suppose is on the roof of this house? On top of everything else, do you suppose they got solar subsidies as well? Cyrus P. Stell, P.E., CEM says: December 5, 2011 at 2:03 pm Very interesting question, Cyrus. It’s particularly important for those items with high capital costs. Let me do a quick back-of-the-envelope calculation here. Suppose fuel price doubles. Lets assume that as a result of the energy price doubling, the price of everything manufactured goes up by say 20%. That seems like a conservative assumption regarding a doubling of fuel. In that case, gas combined cycle fuel costs would go up from about 4¢ to about 8¢. Capital and operating costs would increase by about a half a cent. The detailed calculation shows that the net result of a doubling of fuel prices is that the cost for gas combined cycle goes up by 4.5¢ per kWh, which gives a price of about 11¢ per kWh with doubled fuel costs. Now compare that with solar, where the fuel costs are zero. But all of the capital and operating costs would go up by 20% if fuel price doubles. Solar is currently at 21.5¢ per kWh. A 20% increase in those costs is 4.3¢ per kWh, giving a price for solar electricity of about 26¢ per kWh with doubled fuel costs. So when fuel doubles, gas fueled electricity goes up by 4.5¢, and solar electricity goes up by 4.3¢. Gonna be a long chase … This is why I prefer engineers and business people to scientists for real world questions of this type. w. If anyone you know happens to live in Gainesville, Fla, and you/they obtain their electrical power from the local public utility it’s too late to sign up for the utilities 2012 FIT program (which pays generators $.32 kwh). You can still get into their Net Metering Program- which will pay you about $.115 kwh for energy you send to the grid. These programs (and GRU’s rebate program) are noted in the link below- It was nice to see that GRU is calculating a PV systems output based on the California AC rating method. It will be interesting to see how GRU is going to allocate their increased costs, for the PV they are supporting-- Their Fuel Adjustment per kwh seems like a good place to allocate these costs. Willis writes “That’s exactly why we are discussing levelized costs, Tim, because they avoid all those problems. Please try to follow the discussion. Levelized costs use the net present value of future activities precisely to take care of the issue you are discussing.” No Willis. That report is written in the world where fossil fuels reign and energy supply has large components of costs associated with exploration, extraction, processing and distribution on an ongoing basis. You’re not following my reasoning and instead simply spouting back what the report says. The writers of that report cant possibly take all those factors into account when coming up with a long term energy supply cost of PV solar with any accuracy because eveything is intertwined. Disagree with it if you want, but at least disagree with reasoning as relates to the inadequacies of that report rather than ignoring it entirely. Claude writes ”. ” No Claude because you’re ignoring the fact that the reduction of cost per average Kwh of output due to decreased runnig cost feeds back into the subsequent production cost and nobody but nobody can accurately forcast the impacts of that. Why does not Germany use it’s abundant brown coal and nuclear plants to spin the wind turbines? “Claud also wrote “The bottom line is that centralized, photovoltaic solar, even taken to its most optimistic and theoretical best conversion efficiency” Oh and one more thing, who ever said anything about centralised PV solar? One of the big benefits of PV solar is that it can easily be distributed which benefits in increased redundancy and decreased transmission costs (a less loaded network). TimTheToolMan says: December 5, 2011 at 6:11 pm Thanks, Tim. I thought that I was clear about where I disagreed with you. You talked about the incremental cost to add a single panel. I objected that you had left out the other incremental cost of wires and the upsizing of the alternator and the additional rack to mount the panel and the labor to do all of that … So, where is it that you think I’m not following your reasoning? Finally, I’m not “spouting back” anything. I objected that you had left out the other incremental costs. w. Cyrus: The numbers you published are static numbers, a snapshot taken at a particular point in time. You make an important point. I think, though, that we need two snapshots, and they have different parameters. Snapshot A: I buy solar today and pay a stiff price differential on capital expenses. If prices remain stable for 20 years, I lose. Snapshot B: Twenty years later prices have not remained stable.They have in fact risen considerably, thanks to the combined efforts of the California Greenshirts and Federal voodoo economists increasing the money supply. Coal still retains a huge advantage over solar for new installations, as you and Willis have described, with capital costs for both solar and coal rising proportionally. However, Snapshot B does not include capital costs; these were paid 20 years ago. The fuel costs for solar, essentially zero, have not changed. Thus it is possible that the long-term economics in an inflationary economy could even make solar competitive. (I am, of course, not addressing any of the non-economic problems with solar.) The exasperating thing is that this scenario is exactly what the Greenshirts are trying to achieve with California’s cap and trade scheme. The individual consumer winds up having to protect himself by being the first to invest in something that would make no sense in a stable economy. Well, nuts. I needed a after ‘point in time.’ There’s gotta be a way to show this. You know, a / and an i between a . Mods, help! TimTheToolMan says: December 5, 2011 at 6:26 pm Not only can nobody forecast the impacts of that … I can’t even understand that. I think I understand your words. You say that 1) Solar running costs will decrease (for unknown reasons). 2) This will reduce the cost per kilowatt hour. 3) This in turn will “feed back into the subsequent production cost”. Although I understand the words, I fear that I don’t believe what they say. Why will the running costs (presumably of solar) reduce? In any case, solar running costs are only two cents per kWh, so how can they reduce much? Suppose they dropped by 50% … that’s 1¢ per kWh. And how will that one cent reduction feed back into production cost? w. TimTheToolMan says: December 5, 2011 at 6:31 pm While there are advantages in distributed generation, there are also disadvantages. One-off rooftop installations are generally much more expensive to put in than industrial scale rack mounts. Maintenance on rooftop units generally ranges between “little” and “none”. It means lots of small switchgear, which is more expensive than single large switchgear. Same is true for inverters. Redundancy also means a whole lot more generating units that need to be reliably disconnected from the grid when there is a power outage, to keep from frying the repair personnel. So it’s not all savings, there’s costs as well, and certainly not enough net savings to make up for a 21¢ per kWh price tag. w. Only read first few dozen comments, so this may have been said already: There are so many taxes & subsidies & etc, that the actual, real, costs are pretty much impossible to determine. Last time I read up about this, after a whole lot of digging one item was spectacular clear: wind was by far the most expensive option, greatly in excess even of solar. (More than an order of magnitude above coal.) In other words, I don’t trust your graph. At all. I think it is based on numbers that are flat-out lies. Another point: inverters have an average lifetime of maybe 6 years. Which means that they have to be replaced that often: so integrating them with the panels (25+ year lifespan), would be patiently idiotic. Additionally, the manufacturers of these devices are on record as saying that they do not foresee these lifespan values changing by very much (are inherent to semiconductor components & failure rates). Also, the price-drop curve for the inverters is lagging FAR behind that of the panels… And in any case such curves DO NOT continue forever and ever and ever more downwards: solar-device costs are NOT equivalent to CPU cycles!!! Otherwise we would all be driving 10c BMW’s, pumping 0.001c/gallon gasoline-alternative… none of which, in case the hippies have not noticed, we are not. Willis writes “I objected that you had left out the other incremental cost of wires and the upsizing of the alternator and the additional rack to mount the panel and the labor to do all of that …” Thats fair enough there are certainly initial installation costs involved but not all panels will always require an increase in the inverter capacity if the existing inverter still has capacity and at any rate they’re paid for over much less than the life of the panel. There are concrete examples today actually. I have several friends who have solar PV installations and so far their projected payback period is about 8 years. Probably less if the cost of energy increases as is likely. The solar rebates are decreasing in Australia solar instllations are still affordable and popular. I disagree with your general suggestion that the cost is about as low as its going to get. PV solar is still far from a consumer product in the same way a computer is now. Computers have dropped to a tiny fraction of what they were once worth and the same will happen to any technology that gets mass adoption. Much lower than you think is possible I would suggest. Willis writes “Although I understand the words, I fear that I don’t believe what they say. Why will the running costs (presumably of solar) reduce? ” Because there are virtually no running costs for PV solar. Once installed they pretty much take care of themselves and over time their reliability will improve too. Compare this fundamental feature of the enrergy production to that of oil where there are ongoing costs to produce that oil. Exploration, development, extraction, processing, distribution. All those costs are always with us and always increasing. Now you might argue that you can make more money by “investing” rather than spending on PV solar with a return over many years. And you may well be right but thats irrelevent This isn’t about how you can make the most money, its about how we can best cater to our future energy needs and “investing” sure doesn’t do that. You can “invest” in fossil fuels and sure you’d make more money. Great if thats the goal to make money but its not the path I prefer. .””. Reality has an electric drill in every household and on average its used only a few times ever. People dont mind buying and owning stuff even when they could simply borrow someone else’s drill or hire one. Thats a fact of life. Willis writes “Redundancy also means a whole lot more generating units that need to be reliably disconnected from the grid when there is a power outage, to keep from frying the repair personnel.” You’re dissing redundancy? Distribution companies spend a fortune on redundant feeders and switchgear to manage them. Live line maintenance is common and in the worst case, one only needs to disconnect a bit upstream and downstream of the fault to isolate it. @juanslayton: You have pointed out the inherent weakness of all future casting. We make assumptions, based on the available evidence, and then run the numbers and see how it pans out. My analysis MUST include a life-cycle-cost analysis or it’s worthless. To make that analysis, we decide in advance what we expect future cost increases (or decreases) to look like. EIA projects such numbers, and I use them despite my misgivings (there was another thread, here or somewhere, about how EIA consistently over-estimates future cost of fossil fuels, and under-estimates future costs of “renewables” as defined by legislation). Here’s the thing, though, when we do an engineering analysis we are comparing alternatives. i.e., do I keep what I have, do I buy more of what I have, do I install more of option A (maybe that’s solar PV) do I install more of option B (call that solar thermal) or option C (maybe that’s wind). From that analysis you derive Levelized Costs (gee, where have I heard that term before? Oh, yeah, refer to the title.) Now if we make the wrong assumption about, say, the future costs of electricity, in many cases it won’t matter if we’re comparing something that just uses different amounts of electricity, the same “wrong” costs are in all of the life cycle equations, so it still gives a fairly accurate picture of the best choice. But when we’re comparing different technologies, we have to be a bit more certain of our future-casts. But, you start off assuming no consideration was given to a wrongly forecast future price, and in fact, it was carefully considered. As I originally stated, we assumed 35 years ago that our assumptions (double assumption? I’m leaving it in there) were likely wrong in one direction (future costs of capital would decline while future costs of fossil fuels would rise at some rate) and it turned out we were wrong in the other direction (future costs of fossil fuels in inflation-adjusted $ actually declined for a long time, while the future cost of capital equipment, at least the equipment we were selecting, did not decline, or at least not as much as we had hoped, and might have even risen). So at this point, it’s not just a coin flip to determine who’s most likely right, the figures published by EIA have some analysis behind them, while your hypothesis is just that, a figment of your wishful thinking. Guess who I’d put my money on? Next point, you talk about operating costs of already-installed solar being near zero, and that’s just false. I worked at a place that had made 3 different installations of solar-thermal, and all 3 were deactivated well before the projected end of their expected useful life (the life used in the LIFE-cycle cost analysis) and 1 had been demolished and removed entirely. Why were they inoperative? One thing, well, maybe 2… They got no maintenance, not even drain-down in advance of freezing weather, so much of the tubing burst, and secondly, they were not metering the production of the solar-thermal, so when the question came up, “is it worth it to repair these?” nobody could argue that it was. One of the earlier commenters posted a link to solar installations that had received no maintenance, both solar-PV and solar-thermal, so scroll up. You need to re-evaluate that assumption. I’ve fitted a vacuum tube solar panel to my roof. Did not seek subsidies nor paid the rip-off rates for buying/fitting here in gloomy UK. £500, paid for itself in 18 months two years ago and it looks good for the next 25 years of free hot water for 8 months of the year. Absolute no-brainer. No energy supplier can better that. TimTheToolMan says: December 6, 2011 at 1:09 am None of that makes any difference in the real world, Tim. Sure, once in a while you’ll have spare capacity in the inverter, but in general it was originally sized to match the panels so you can’t just add panels. Also, whether the inverters are ” for over much less than the life of the panel” is immaterial. You still have to pay for them, right? How is that a “concrete example” of increasing panels? All that proves is that the Australian taxpayer is subsidizing your friends’ green fantasies. Perhaps that is because I never made that suggestion. I said that solar could go lower, and I estimated how low it was likely to go. I showed that even if panel pricess dropped through the floor PV would still be uneconomical … where were you when that discussion was going on? You should refrain from suggesting if that’s the best you can do. I estimated above the likely lowest cost you’ll find a complete system if panels get really, really cheap. It’s at about 18¢ per kWh. It will not drop to a “tiny fraction” of that cost, any more than cars will drop to a “tiny fraction” of their current cost. In fact, you will be lucky to get down to that cost, because the panels are only about 30% of the total cost at present. That means that if panels were free, the system would still be twice as expensive as fossil fuel. w. TimTheToolMan says: December 6, 2011 at 1:18 am Running costs for solar are currently about 2¢ per kWh. You said they would reduce. I asked how they could reduce much when they are so small. Your answer is that they will reduce because they are so small … say what? They will go up if other costs go up, including fuel costs. I see the reason for your confusion. You are comparing the running costs of a solar electric system with the extraction, refining, and distribution costs for oil. You seem to be under the impression that oil prices have always gone up. Nothing could be further from the truth. We pay about the same (in constant dollars) now for fuel as we did in 1950. But in any case, none of this reduces solar costs as you have claimed. It’s the other way around—as I showed above, if fuel prices rise, the cost of solar installations rises as well … and for a doubling of fuel prices, the increase in costs is about the same for both systems. I would never argue that. I would argue that you can lose more money on solar than on any electricity generation scheme except offshore wind and solar thermal. That’s a very different claim. It’s about the cost of electricity. You want to make it some kind of noble quest for the holy grail of our “future energy needs” or something. For the rest of us, it’s about cost. We know that our future energy needs will be met by our future citizens in some future way, and it may have nothing at all to do with either solar or fossil fuels. You seem rather attached to this point, that somehow fossil fuels are not the “path you prefer”. I have no problem with that. What I do have a problem with is paying for your preferred path. If you want to take it, go ahead. But sticking your hand into my pocket to pay for it? Why should I pay for your preferences, Tim? w. So, Willis, I’m skeptical of the safety of fracking and I mispelled aquifer and I’d rather see coal burned for electricity than gas. Y’all are tough on skeptics here. TimTheToolMan says: December 6, 2011 at 3:14 am Actually, that statement about computers, as far as anyone can determine, is an urban legend … kinda like your claim that somehow an imaginary statement falsely ascribed to Tom Watson has relevance to your argument. How are electric drills and Thomas J. Watson relevant to my statement, which is that a fifty kilowatt inverter is going to cost you less than ten five-kilowatt inverters? My statement is clearly true, and you respond with urban legends … kind of symbolic of the whole discussion. For years they didn’t allow direct connection of your own generation gear (aka rooftop solar) to the grid because it was too dangerous to the repair people. Your claim seems to be that it was never a danger … riiiight … No, I’m not “dissing redundancy”, Tim. I am a realist who knows that there are problems that come along with every solution, including redundancy. For generator redundancy, one of the problems is disconnection of all of the generators in cases of system failure. Claiming I’m “dissing redundancy” doesn’t make that problem magically vanish. And having a 30kW rooftop solar system still connected to a down system is a very bad thing. Power goes down … except for one holdout solar system who didn’t get the word. PGE shuts off the mains, but the lines are still live. The lines will remain live, and very dangerously so, until somebody shows up to cut them off. How is that not a problem, especially in an emergency where help may not arrive for a couple days? And contrary to your claim, you need to do more than “disconnect a bit upstream and downstream”, because that still leaves power on the the rest of the grid, and when the grid is down, that’s dangerous. Next, in many grids, there is no well-defined “upstream” or “downstream”, particularly if the grid is fed from many, many points as you advocate. Finally, what about the homeowner who decides to reconnect his system to the grid during a power outage, or does so accidentally? So your claims about how the generator disconnects don’t matter are an urban legend as well, Tim. 100% reliable disconnection of live generation equipment is a real concern for anyone operating a power grid. And my more general point is true. There are advantages to redundancy, but there are also problems with redundancy that you are not taking into consideration. w. TimTheToolMan says: December 6, 2011 at 3:14 am .””.” Here, Tim has a point. We’ve only just started to seriously mass-produce inverters; there’s a lot of room for driving down the production cost. Compared to say, motor electronics, which are produced nearly 100% automated – a typical assembly line produces 5,000 of them or so in an 8 hour shift, and a factory has maybe 10 such lines – inverters are still large boxes with a lot of manually connected cables. This will change when we use one small inverter per module; they become as small and as plentiful as motor electronics boxes, in the millions of pieces range, and the same automation will be used for mass production. Shortly thereafter they might become solid state modules, and later one chip solutions. I’m fantasizing here, but that’s what usually happens when the numbers are scaled up. ” Reality has an electric drill in every household and on average its used only a few times ever. ” Tim, you shoot yourself in the foot here – that’s exactly not an example for high efficiency ;-). mike g says: December 6, 2011 at 11:34 am You can pretend that was the problem if it helps you to sleep, Mike. I thought I was quite clear, but if not, the problem was that you were waving your hands at the web and telling me to go google for the information to support your argument. Not my argument. Yours. I don’t do that. I don’t go on your quests for you. If you have a point, you’ll have to make it and support it with citations yourself. We are not “tough on skeptics” here. We’re tough on people who want us to do their work for them. Me, I’m still waiting for what I requested, a statement of your specific claims and the citations to back them up. Until you provide those, you are just whining about how you are being treated … and that don’t impress me much. You want better treatment? Then tell us exactly where fracking did something bad, when it did it, what happened, who was involved, and provide us with the details of the incident. Because saying “I’m skeptical of the safety of fracking” is meaningless without the details . w. Willis Eschenbach says: December 6, 2011 at 11:46 am “No, I’m not “dissing redundancy”, Tim. I am a realist who knows that there are problems that come along with every solution, including redundancy. For generator redundancy, one of the problems is disconnection of all of the generators in cases of system failure.” Willis, all solar inverters used in Germany need to switch themselves off as soon as the grid frequency rises above a certain threshold – because this frequency signalizes overload. Similarly, a lot of emergency conditions are already encoded in the inverters; they have a digital signal processor that does about 15,000 cycles a second and checks these conditions. Most inverters will only feed in when an outside grid exists; they never build up a grid by themselves! Disconnecting the grid suffices to make them all shutdown by themselves. On December 3, 2011 at 8:03 pm, Willis Eschenbach says: “why do you have to be so snarly and ugly and nasty? All it does is make you look like a vicious, vindictive little man.” Pot, kettle? Ducky12 says: December 6, 2011 at 12:02 am So if I have this straight, you gave up easily on the actual analysis because there were “so many taxes”, you say it’s “pretty much impossible”, and you call the numbers “flat out lies” … Let’s see. Here’s the order of battle. On one side we have the Energy Information Agency, who is actually trying to put out real numbers, and generally acknowledged and well respected in the industry for doing so. They have spent hundreds and hundreds of hours, they didn’t complain that there were “so many taxes”, they calculated the numbers as best they knew how. And other the other side we have a random internet poster named “Ducky12″ who says that the math and the work is too tough for him, he can’t figure it out, so it all must be lies … Ducky, despite the trenchant, biting nature of your keenly incisive analysis of the situation, I fear I’m going to go with the EIA just this once. Yes, I know there are some difficulties with their numbers, and I have referred to some of those problems above. But overall, they are the best numbers we have, and as long as we keep an open mind that they are best estimates, and that any one of them may be a bit high or low, they are quite useful to us. Throwing them out would be the act of a petulant child. Instead, we work with what we have, and simply include the uncertainties in our analysis. w. Louise says: December 6, 2011 at 12:00 pm Why pot and kettle, unless you are referring to yourself? Are you revealing your secrets? Louise, have you been fooling us all, have you been sneaking off to where Dave posts, not to discuss the issues, but simply to be a troll and muddy the waters? Do you follow Dave around the web, looking to attack him wherever and whenever he posts, and getting all snarly and ugly and nasty, the same way that Dave does to me? I know I don’t do that to Dave, which means it can’t be me you are referring to. I simply call Dave a vicious, vindictive little man when he follows me around and acts like one. Other times, when he wants to talk science, I do that instead. You have a problem with that? Sorry, but that’s how people respond to vicious, vindictive little men. So if there is a pot to his kettle, either that’s you, my dear, or you’ll have to clarify to whom you are referring. w. PS—Here’s Dave’s intro, his very first comment to me in this thread, his opening salvo before I had said a word to him. It’s typical vintage Springer: Like I said, he is a vicious, vindictive little man. You don’t like me calling him that? Sorry, Louise, I call them like I see them. It’s an enduring fault of mine. PPS- Here is an example upstream of where I answered Dave’s scientific point. I did so to once again check if he is actually interested in a discussion. I simply asked for a citation for his claims. I got nothing in return, which is typical of the way Dave deals with scientific questions. Louise, your defense of the man speaks very well of your heart … but very poorly of your mind. DirkH says: December 6, 2011 at 11:59 am … My point was not that the problems of distributed generation could not be solved, at least most of the time. Like any problem (except for finding the value of the “climate sensitivity”), they can generally be solved given enough materials, time, and money. I was responding to Tim, who seemed to think that there were only advantages to distributed generation, and no disadvantages or hidden costs in distributed generation. In fact, one of the disadvantages is that you need (as you point out above) special switchgear to disconnect each solar rooftop installation. And that special switchgear is neither free nor 100% reliable … so in fact there are disadvantages and additional costs from distributed generation, not just advantages and no additional costs as Tim was blithely claiming. w. DirkH says: December 6, 2011 at 11:51 am No, Tim doesn’t have a point. he has a manufactured, false quote that doesn’t support his claim, and that seems to have fooled you as well. My point was that ten 5 kW inverters will cost more than one 50 kW alternator. You have not brought up anything that shows my point wrong. Yes, prices are falling on electronic gear. Yes, they could come down more. But that’s not what Tim was claiming. He was claiming that there were lots of benefits to distributed power … and no disadvantages. I pointed out that one disadvantage is that many small things cost more than one big thing, inverters being a prime example And now, although you obvious don’t understanding what the point of the discussion was, you want to jump in and reveal to me the big secret, the thing you think I don’t understand, that inverter costs are falling?? Ummm … well … thanks, but I’ve known that for years, DirkH, and it makes no difference to the topic under discussion. Dirk, you seem like a real smart guy, but you are not following the story. Go back and read the interactions between Tim and myself. The discussion is not about whether inverter costs will fall, as you seem to think. It is about whether there are additional costs and disadvantages to a distributed solar generation system, compared to having all of the solar in one place. w. Willis Eschenbach says: December 6, 2011 at 12:40 pm …” I’d much rather trust a system designed under safety-critical considerations than anything else, because that’s the best bet I can make. As an additional precaution, gloves can’t harm :-) “In fact, one of the disadvantages is that you need (as you point out above) special switchgear to disconnect each solar rooftop installation.” That’s exactly one of the beauties of one inverter per module – the inverter semiconductors do the disconnect. When you have a large inverter for a string of modules, you get a DC voltage of about 800 V; that can be nasty. Not so with the microinverters. You have the output DC of one module, that’s 60V or so. (I’m guessing). Even if the the emergency switchoff fails, the voltages are not that dangerous. At the moment, efficiency lags behind big inverters and total costs are higher. We will see. I’m not trying to advertise anything, at the moment I have no business interests there. Timthetoolman has a few statements- 1) “Because there are virtually no running costs for PV solar. Once installed they pretty much take care of themselves and over time their reliability will improve too.” 2) “I disagree with your general suggestion that the cost is about as low as its going to get.” Tim, In response to statement 1: My PV system has been in service for 5.5 years. In that time I have had my inverter serviced once (software was incorrect in how it calculated kwh’s- it read low by about 25%)- the upgraded software then read about 10% high per a separate kwh meter I had installed just before my PG&E E-7 net meter. After about 2 years of service my inverter started reporting negative values for one of the attribute it reports (instantaneous wattage). Rather then try to figure out what was wrong with my original inverter the unit was replaced. The new inverter has worked fairly well- it reports my kwh output about 4% high per my secondary kwh meter. The estimated mean time to failure for inverters is around 10 to 15 years, so any cost calculation needs to take this into account. Additionally, the efficiency of the inverter is not going to improve over time. If I don’t keep a fig tree trimmed, that is located just south of a couple strings of my panels, my output drops by about 20%. Hence I have some ongoing maintenance for my system to operate at it’s rated POTENTIAL max output. Over the years I have found that without cleaning my panels (especially in the dry, dusty summer months) my overall output will drop between 8 and 10%. My weekly preventative maintenance is rather straightforward- a rinse with water from my garden hose with a fairly strong stream of water- geese fly overhead occasionally and it takes a bit of water pressure to remove their droppings from my panels). As far as reliability goes each manufacturer of panels is required (in CA anyway) to limit (warranty) the degradation of their potential max output (STS rating) over the expected life of the panels. For my panels an expected degradation in STS max rating was something like a 20% reduction in STS max output over time (time being 20 or 30 years for my panels) of the warranty. In response to statement 2- “I disagree with your general suggestion that the cost is about as low as its going to get.” I concur with Willis on this one but with a change in the word cost to price for an installed residential PV system. As the balance of system (copper wire, inverter, aluminum railing system and installation labor, shipping) costs become a larger part of the overall system costs a 5 to 10% improvement in the cost of the panels are going hit the law of diminishing returns from a total price to purchase a self generation option. The price I paid to have my 6.12 Kw system installed in 2006 was $1.10 a Kw. Today the labor and misc materials (some cu wire, shut off switches) costs are between $1.30 and a $1.50 a Kw out here in CA to put a 6.12 Kw system in. The price paid by the wholesaler who I bought my panels from has gone down with the drop in panels costs. Their costs for the Al railing and their cost for the inverter haven’t come down. Their costs for shipping (primarily fuel costs related) have gone through the roof so free shipping is no longer included in the purchase of the hardware components of a PV system. The best price (before rebates, and tax credits) to have a mid sized investment grade residential PV system (4 to 10 kw) installed are likely never going to get below $3.00 to $4.00 (CEC AC rated) watt. Under my version of a best case scenario for residential PVgeneration (lets use the VERY optimistic $3.00 (installed watt cost)* 5.22 CEC Kw rating (the AC rating of my PV system)= $15,660.00 which yields a total yearly output of 9300 kwh- for a 6.12 kw STS rated system). The upfront costs in this best future scenario is $15.6K for a 5.22 kw cec rated residential system and it will yield 9300 kwh a year DirkH says: December 6, 2011 at 1:04 pm So clearly, the answer to my question that you are unwilling to give is that, no, you would not be willing to bet your life that the disconnects would work. Next time, just answer the question, it makes you look shady and evasive when you answer everything but what was asked … Oh, they do, do they? And you can point to a system where this is installed and working, I suppose, but you just forgot to do so? As far as I know, you are talking about a fantasy, a system with an inverter in each panel, which has never been commercially available. But somehow, you didn’t let people know that you were talking about imaginary panels, you acted like you were describing some real system that had been installed and tested. Bad Dirk, no cookies … As you said, Dirk, you are guessing. You don’t know if anything you are claimed is a fact, because there is no such system yet built, just your fantasies about such a system. Not only that, but that wasn’t the point of the discussion you jumped in the middle of. We were discussing if ten 5 kW inverters cost more than one 50 kW inverter. Somehow, despite my pointing it out, you haven’t gotten back to that either. You should start over, talk about real things that actually exist and that you know something about, and follow the discussion. Your guesses about random panel designs that have never been manufactured are meaningless. w. Willis, you’re really having a bad day; I’ll leave after this comment. Nobody forced you to answer my comment; so blaming me for saying “Tim has a point there” with “We were discussing whether this kind of inverter is cheaper than that”, well, that’s just silly. No offense; don’t work yourself up. Bye. “Released in 1993, Mastervolt’s Sunmaster 130S was the first true micro-inverter.” Looks pretty real to me: kakatoa says: December 6, 2011 at 1:17 pm Thanks, kadaka. 9300 kWh per year from a 6.12 kW rated system gives a specific yield of 9300/6.12 = about 1,500 kWh annually per kW of installed power. That’s a little better specific yield than here in the San Francisco Bay area, I assume that either you are further south or are using theoretical numbers. So if we assume a 30 year lifespan for a levelized cost comparison, and two cents per kWh for running costs, your imaginary “best-case” system is costing out at $15,660 divided by (9,300 kWh/yr times 30 years). That gives us about 8¢/kWh for the power. However, I think your estimates are off for a couple of reasons. At least around here, prices are about $7 / kWh of installed DC power, which in turn is less than the amount of AC power you generate. This estimate of our local cost here (SF Bay) is borne out by the LA times, which comments: This means that the current cost for your rooftop power system would be about $7.50 * 6.12 / (9300*30)+2¢ for operations, or 19¢ per kWh. I doubt that it will ever be cut by more than half as you suggest, because that would mean that every component, from the panels and racks to the inverters and the labor, would have to see a more than 50% drop … very doubtful on my planet at least. It will come down, but not that far. So you are right that yours is a “VERY optimistic” analysis. Either your analysis or mine comes to the same conclusion, however, which is that solar will not cost less than fossil fuel any time in the foreseeable future. w. The chart shows 16 electrical energy sources. The costs of production do not include externalities, so they constitute a rough guide on real costs of production. One poster noted that solar hot water, by comparison, in effect delivers energy at a much lower cost, yet this source of energy is omitted. I also note that a British study has showed that the cost of reducing electricity demand varies from zero (in the case of turning down a thermostat, for example) to less than 1 cent per kw for many simple technologies like lagging of hot water pipes, but almost always much less than the unit cost of supply. This endless and often bitter dogfight over preferred energy supply technologies is interesting, but it has very limited horizons. DirkH says: December 6, 2011 at 1:58 pm Do they really do that where you grew up? Do they respond to someones reasoned discourse by saying that the person is “just silly”, and then add “No offense, don’t work yourself up.” No offense? You insult someone, and then you think saying “No offense” somehow makes your comment socially acceptable??? Not where I grew up. We stand behind our words, we don’t say things to offend somebody and then say “No offense”. In addition, you did come into the middle of a discussion to support one side, by claiming that Tim had a point. Then, instead of defending Tim’s point, you wandered into a whole other topic. And when I point that out that you claimed to be defending Tim’s point but you didn’t understand or even discuss the point we were talking about … you get into a snit, take your toys, and go home. Gotta love the grown-up type behavior. You go on to say: From your reference: OK, so someone does actually make what they call “micro-inverters”. I thought you were discussing inverters that were a part of and integral to the solar panels, not some add-on type product. My bad. However, your reference goes on to say: Since that was my point, since that was the issue that Tim and I were discussing, that more inverters would indeed be more expensive, you have proved it very neatly. So I will follow your lead, and say “… don’t work yourself up. Bye.” w. PS—As I said before, you do seem like a sharp guy. If you want to get all huffy and go home because I had the insufferable gall to actually disagree with you, you are certainly free to do so. However, do you think that increases or decreases the chances of you being seen as a serious player? Do you think that increases or decreases the chances that your ideas will be believed? Solar power may be useful as an auxiliary power source, but eventually carbon-power is not going to be an option. (Bio-solar carbon excepted.) I am not prepared to say when that might be, but it does look like we are now consuming more carbon-power than we are discovering, at least in the case of petroleum. I look at this as a question of whether energy from the sun can replace exhausted carbon-power in all uses. How much of the earth’s surface must be set aside so that each person can collect all the energy needed for heating, transportation, and feeding from the sun. That presumes the manufacture of synthetic transportation fuels. I am guessing that the average recoverable solar energy is less than 100 watts per square yard. We do know that back in 1880 we had a lifestyle model and total global population that did not depend on carbon-power to the extent we do today. Willis writes “Running costs for solar are currently about 2¢ per kWh. You said they would reduce. ” Come on Willis, get a grip. From your original diagram, do you seriously think that running costs of Coal at 3c vs running costs of Solar PV at 2c is correct? WTF Wills? “I was responding to Tim, who seemed to think that there were only advantages to distributed generation, and no disadvantages or hidden costs in distributed generation. ” You’re the one who brought up the disadvantages of many generators connected to the grid without mentioning the fact the added redundancy was also a good thing. Chris Harries says: December 6, 2011 at 2:05 pm Thanks, Chris. That’s why the chart refers to “Generation Resources”, because it is discussing electrical generation. In fact, that’s the subject of this thread. Which is why hot water is not on the chart. It is true that in places where energy is routinely wasted, there is often “low-hanging fruit”, places where conservation is a cheap, hugely efficient no-brainer. In the rest of the world, however, this is not true. People at the lower end of the economic spectrum (AKA most people on the planet) have very little slack in any of their budgets, energy or otherwise. They picked the low-hanging fruit generations ago. Finally, while your point is well taken that conservation is generally cheaper than generation, here in California our power prices are going through the roof because of green fantasies. You’ll excuse me if I find that to be a separate and independent topic that is as deserving of discussion as is conservation. So what I do on WUWT is pick and discuss one thing at a time (as much as is possible in this interconnected world). This does not mean the other things are not important. It also does not make the discussion have “limited horizons”. Our horizons are the same … we’re just discussing things one at a time. This thread is about electrical generation. Thanks, w. TimTheToolMan says: December 6, 2011 at 4:21 pm What on earth does coal have to do with your claim that solar costs will reduce? Let’s get that out of the way before you run off to discuss coal. As to the coal costs, coal is very cheap per kWh. Why do you think it is used all over the planet as an energy source? Why do you think the Chinese are building one coal fired plant a week? I suspect that the EIA figures are not too far wrong. It’s a mature technology, there’s got to be tons of data on the costs of coal. So yes, I do think they are definitely in the ballpark for coal. Now, could we return to your claim that solar running costs will reduce? w. Spector says: December 6, 2011 at 2:46’s, the US is about to become a net exporter of refined petroleum products. Your claim is simply wrong. Since we also have energy available from nuclear fission (and perhaps fusion as well), I don’t see why you are limiting your analysis to the sun. In any case, 100 watts per sq. yard (or square metre) as a total recoverable solar yield seems high. Let me think about that. Louisville, Kentucky is in the mid-latitudes, around 40 N. Total insolation is about 4 kWh / m2 / day. This gives 4000 watts divided by the number of hours in a day, which gives an instantaneous average energy produced of about 170 watts / square metre over the 24 hours. A typical solar system might have an efficiency of 10% or so, which means that currently we could harvest about 17 w/m2 in Louisville. Will we get to where we are six times that efficient, to get to your 100 w/m2 figure? Perhaps … but that’s a ways away. If you wish to return to a “lifestyle” where kids routinely died because there were no antibiotics, that’s easily done. Just move to rural Africa. But don’t expect me to clap and join in the move, thanks. w. TimTheToolMan says: December 6, 2011 at 4:33 pm I brought it up??? I didn’t mention that there were advantages to redundancy??? Do your homework before you start up with false accusations. Here is my first mention of the question, in response to a claim of yours: That was our introduction to the discussion of distributed generation. Note that you, not I, brought up only one side, the benefits. I was responding to that lack of balance. Note that in my response it was I, not you, who stated that there were both advantages and disavantages. In other words, you are reduced to simply making things up … and not only that, you accuse me of what you did. Check before posting, my friend, it will keep you from such transparent and easily avoided errors. w. Evening Willis, The 9300 kwh per year I noted is my actual 6.12 kw STS rated (5.22 CEC rated AC) output for my system- measured using the separate kwh meter I had installed after my problems with the non accurate numbers from my original inverter. I reside up in the foothills (out of the fog and most of the time out of the snow) south of Highway 50 at a 2400 elevation. My panels are roof mounted facing south (the building were built with the roofs this way- I didn’t have anything to do with it) with 12 of them at a 30+/- % slope and the other 24 at a slope of 16%. Back when I was trying to find a way to pay PG&E less then I did in 2005 (we had a yearly electrical bill of $3600 and we used 2200 kwh from PG&E in 2005) I looked into self generation as an option. The local airport had loads of wind speed data- the data said the average and SD of the wind speed in my area were not appropriate for any self generation wind option. I thought of just telling my 80 year old parents (who were living with us at the time) that they couldn’t turn on the air conditioner or heaters….. but that didn’t seem like a good plan. Back in late 2005 and early 2006 the PV installation industry wasn’t to developed- It was near to impossible to get anyone to come to my homestead- we live on 11 acres- to quote on putting some PV on a couple of my buildings. My make/buy decision was made a bit more difficult by the three firms that finally came out to the homestead to give me a quote. Yes, the sales representatives reminded me of a less then enjoyable used car dealer experiences. To make a long story shorter I ended up converting all the hardware (and installation) quotes into AC output per year as that way I could compare apples to apples. I got quotes all over the place price and output wise. And yes their quotes came out close the price noted in the LA Times article. I ended up being the general contractor for the project- I selected the hardware and location for the panels- and selected a reputable installer to put the hardware up (I would of lost 20% of my rebate if I had done the install on my own by the way) to save some money on the system. My wholesale hardware sourced system, before rebates and tax credits, ended up being $7189 Kw (DC) in 2006. I did a quick check at the wholesale PV supplier who I got my hardware from and to find out what wholesale PV system pricing would be for a system that matches the output of mine- using the same manufacture for the panels. Per their web site it would cost $20808 to get the same 6.12 kw (STS rated) output without tax with free shipping. So it looks like one could put a similar system in place in the Bay Area today for: hardware= $20808 +1768.68 (tax)= 22576.68 +7956 (install costs at $1.3 per kw)= $30532/5.22 kw CEC rated= $5900.00/ kw ac which is a bit less then your example. (BIG NOTE TO THE Moderators- your may or may not want to include the reference to the wholesaler I used back in 2006- ). My younger brother is rather handy (which comes in handy as he actually knows what he is doing when it comes to down to Mechanical and Electrical engineering items- his degrees). If I was going to be a general contractor for a PV system in the future I could likely convince John to visit for a week and cut the install costs down by half. These days the federal tax credit for going with a PV system (green?- well that is another question as my power actually comes from a small hydro facility that feeds my distribution line) is 30% (it was limited to $2k back in 2006) so the final out of pocket cost for a PV system, installed by a contractor (vs. my brother) would be $5900.00- (.3*5900)= $4130 per AC kw. in 2011. I think the PG&E rebate amount for PV is down to something like 30 cents a kw which you can get from PG&E if your don’t install the system yourself and I didn’t include this in the cost side of the equation as I don’t know what the permitting costs are these days (likely a lot more then the $250.00 I paid back in 2006). On the benefit side of things in CA one can leverage PV (an enabling technology) to risk going with a time of use meter (E-7 in my case) which reduces the cost of a kwh to $.075 for baseline power from PG&E during off peak times. Additionally if you can really manage your load it allows you to get a .30 credit for kwh’s sent to the grid at peak times in the summer. I concur that for base load and reliable (i.e. no intermittency and scalable) peaking electrical generation natural gas fired plants are the way to go. RE: Willis Eschenbach says: (December 6, 2011 at 5:22′s, the US is about to become a net exporter of refined petroleum products. Your claim is simply wrong. I base this on Figure 3 published in David Archibald’s article, “Peak Oil – now for the downslope” and his reputation as a petroleum expert. Jeff Rubin, *former* CIBC chief economist, has been writing in the Globe and Mail that the US debt situation makes it likely that China and India will be the primary customers for Canadian oil. ”Since we also have energy available from nuclear fission (and perhaps fusion as well), I don’t see why you are limiting your analysis to the sun.” I am posing this as a response to those who insist that we all must depend on traditional ‘renewable’ energy resources for the good of the planet. ”In any case, 100 watts per sq. yard (or square metre) as a total recoverable solar yield seems high. Let me think about that. That was an absolute upper limit guess as to what might be obtained as an annual average including the effects of cloud and hours of daylight. It may well be much lower as you suggest. That greatly increases the footprint of the solar farm that would be required to replace carbon-power. [We do know that back in 1880 we had a lifestyle model and total global population that did not depend on carbon-power to the extent we do today.] If you wish to return to a “lifestyle” where kids routinely died because there were no antibiotics, that’s easily done. Just move to rural Africa. But don’t expect me to clap and join in the move, thanks. I think something like that is just what might happen if we are all eventually constrained to the use of traditional ‘renewable’ energy resources. Tim (the Tool Man), here’s an article about the advantage that you claim for distributed systems, that it leads to “decreased transmission costs (a less loaded network)”. Here’s the reality: Source I did love this explanation. TimtheToolMan, think about this one (emphasis mine) … This is the problem with all these fantasies. Nobody thinks them all the way through. They figure if one of something is good, then a hundred of them will be great … until the system starts overloading. w. RE: Willis Eschenbach says: (December 6, 2011 at 5:22 pm) “A typical solar system might have an efficiency of 10% or so, which means that currently we could harvest about 17 w/m2 in Louisville.” According to one reference, (T Boon Pickens) we are currently harvesting about 85 million barrels of petroleum a day. Each barrel has the energy equivalent of about 1.7 million watt-hours. Thus we are extracting 145 terawatt-hours a day or a sustained global energy flow of about six terawatts. A solar farm required to produce that much energy by your figure would be about 354 square gigameters or about 600 kilometers on a side. The installation cost, maintenance, and environmental consequences of such an array are indicative of the dark future of the exclusive use of solar power. It would be ironic if an array that large influenced the climate. Keeping Warm The Aussies need a stern talking to by TimtheToolMan. They clearly haven’t realized that distributed systems are so much superior to centralized systems, or as Tim put it … Here’s how Tim’s “less loaded network” is playing out in Australia (emphasis mine) Now, what do you think will happen because of this? Costs go up. There’s a surprise. And remember, dear friends, that these are the kinds of costs of renewable energy that are NOT included in Figure 1 above. The article continues: The joys of distributed solar electricity … w.. ————– That’s correct but only if the other costs do in fact stay fixed. The cookie cutter nature of PV means that the whole system is subject to economies of scale. It is plausible that you could have factories building entire systems on assembly lines like cars. Given market barriers this might not happen in the USA. But in China they need to build a lot of power sources of all kinds, but especially ones that prevent them from choking to death on SO2 aerosols from coal burning. So while I am not an enthusiast for PV those costings from Willis could become out of date very quickly. LazyTeenager says: December 9, 2011 at 9:29 pm Hey, Lazy, good to hear from you. You are right about the “cookie-cutter nature” of large photovoltaic installations. I think, however, that you are drawing the wrong conclusions from that fact. The problem is that PV is not new technology by any means. It has been built and pushed and subsidized for thirty years or more. As a result, the low-hanging fruit in terms of economies of scale have already been picked. You won’t see the cost of the racks, for example, dropping by much. Why not? Because there are a variety of manufacturers, and there is heavy cost pressure on them already. They are popping them out as cheaply and quickly as they can, and they’ve been doing it for a while. Regarding China, they are building nuclear and coal. They do not have our inhibitions about breeder reactors, so they will have no shortage of nuclear fuel. As their economy improves, they will require the type of pollution controls on their coal plants that we have in the US. Those two will fix the pollution problems you mention above. In the meantime, the Chinese will happily make solar cells for export sale to the West. And as someone pointed out above, they hope to be paid by the West to retire their coal plants in favor of nuclear when their coal generation systems start wearing out. All the best, w. From an editorial in the Washington Times: The oil figures seem high, but the rest sounds about right. w..” If all nuclear power options were deemed to be unsustainable or impractical or just ‘politically incorrect’ then solar power or solar derived power is all that remains. I do not believe this energy source, alone, could sustain our current population, once the Earth’s sources of carbon power are exhausted. I believe we should be looking for a practical replacement energy source now, before we have to start dealing with the consequences of reduced energy availability. As far as ‘taxfree’ goes, I expect you would have to pay property tax on the value of your solar power installations and there may well be a pollution impact in gathering the raw materials and the manufacture of your solar panels. You might also have to pay a monthly maintenance fee to keep them working efficiently.
http://wattsupwiththat.com/2011/12/03/the-dark-future-of-solar-electricity/
CC-MAIN-2015-18
refinedweb
47,192
62.07
![endif]--> Arduino <![endif]--> Buy Store Distributors Download Products Learning Getting started Examples Playground Reference Support Forum Advanced Search | Arduino Forum :: Members :: digimate Show Posts Pages: [ 1 ] 2 3 4 1 Using Arduino / Motors, Mechanics, and Power / Re: Help identifying some salvaged motors, wiring, voltage, etc ? on: April 15, 2013, 07:05:02 am Managed to find a page that details the typical stepper found in old 5.25" drives here - It even has the wiring colour codes and typical current. That's one of them sorted 2 Using Arduino / Motors, Mechanics, and Power / Re: Help identifying some salvaged motors, wiring, voltage, etc ? on: April 15, 2013, 01:42:36 am Quote from: semicolo on April 14, 2013, 07:50:17 pm The parts coming from the floppy are likely to use 12V or 5V, minus the driver drop. So they don't usually regulate the 12V down to something lower? OK, good. Quote from: semicolo on April 14, 2013, 07:50:17 pm The way I guesstimate the voltage of unknown steppers is apply a voltage to 2 windings out of the 4 available and wait for the temperature rise. Then I slowly increase voltage until the temperature stabilizes around 70C (some steppers can take more but without a datasheet it's best to keep a safety margin). You can probably use this technique for the solenoid/dc motor, maybe with lower temperature rise. Good idea to test via temperature, thanks. I am confused about the stepper, I thought they had 2 windings, not 4 - also the big one has 5 wires to add to the confusion. 3 Using Arduino / Motors, Mechanics, and Power / Help identifying some salvaged motors, wiring, voltage, etc ? on: April 11, 2013, 08:41:56 pm I would appreciate an help you may be able to give me on making use of these bits and pieces, particularly wiring and voltage identification. From an old 5.25" floppy drive... 1. Minebea Co.Ltd Stepper motor (head positioning) 17PS-C006-01, 5 wires white/brown/red/black/green 2. Beuhler DC motor (main drive), 2 wires from top blue/red, 2 wires from bottom (near the pulley) green/yellow 3. Solenoid (head clamp) (made in mexico) 371-30, wiring is easy - just 2 metal tags, but what voltage? From a CDROM drive... 1. Small unbranded stepper with worm drive from laser positioning in CD rom drive (04X17R290), wiring is flex. cct strip, 4 conductors from junk box... 1. Small unbranded DC motor VD 030Y22 4 Using Arduino / Motors, Mechanics, and Power / Re: Motor shaft sizes - do standard sizes exist? on: April 05, 2013, 07:17:06 pm Quote from: MarkT on April 05, 2013, 06:31:11 pm You can get gears/pulleys for a smaller shaft and drill/ream them out. 4.1mm sounds unlikely - micrometer or calliper guage is the way to measure. They measure 4.18 to 4.2 with a digital calliper (depending on where on the shaft I put it). Must be a weird size. Good idea on drilling/reaming smaller sizes - thanks. 5 Using Arduino / Motors, Mechanics, and Power / Re: Motor shaft sizes - do standard sizes exist? on: April 05, 2013, 06:09:21 pm Quote from: kf2qd on April 05, 2013, 12:43:53 pm Those) Ah, sounds messy . Sounds like you just have to adapt where you can't get the exact right size. Thanks for the help. 6 Using Arduino / Motors, Mechanics, and Power / Motor shaft sizes - do standard sizes exist? on: April 05, 2013, 04:30:39 am Just starting out using a few DC motors, and some servos, stepper motors etc., and I was wondering if there is any standard sizes for the motor shafts so that standard size gears/pulleys etc can be purchased off the shelf to fit. For example, I have 3 very nice 12VDC Pittman motors from an old computer data storage device. Their shafts measure about 4.1 or 4.2mm. Am I going to be able to find any pulleys/gears, etc to fit them? Related question, how do you couple mechanics to your motors usually - a flexible coupling, belt/pulley drive, direct? Do you make them yourself? I had a suggestion from a guy who sells Lego gears and stuff, that people often use a piece of tubing, such as used for air hose in the aquarium trade. 7 Using Arduino / Displays / Re: M2TKLIB Hello World with LiquidCrystal_I2C on: February 14, 2013, 12:59:22 am Yes, was just writing this when you posted - now the M2TKLIB Hello World example works, modified as below. Changes are to include <Wire.h>, to include <LiquidCrystal_I2C.h> instead of <LiquidCrystal.h>, and to set it up using "LiquidCrystal_I2C lcd ( 0x3F,2,1,0, 4,5,6,7,3,POSITIVE);" Code: #include <Wire.h> #include <LiquidCrystal_I2C.h> #include "M2tk.h" #include "utility/m2ghnlc.h" LiquidCrystal_I2C lcd ( 0x3F,2,1,0, 4,5,6,7,3,POSITIVE); M2_LABEL(hello_world_label, NULL, "Hello World!"); M2tk m2(&hello_world_label, NULL, NULL, m2_gh_nlc); void setup() { m2_SetNewLiquidCrystal(&lcd, 16, 2); } void loop() { m2.draw(); delay(500); } Thanks for ALL the help ! 8 Using Arduino / Displays / Re: M2TKLIB Hello World with LiquidCrystal_I2C on: February 14, 2013, 12:41:23 am Another step forward ! Figured out the constructor numbers are bit numbers, so I had to follow traces on the backpack board to see where the pfc8574 pins were wired to on the LCD display proper. Then translate those to bit numbers. For anyone else with a sainsmart LCD display with their "lcd2004" I2C adapter, the following seems to work. Code: LiquidCrystal_I2C lcd ( 0x3F,2,1,0, 4,5,6,7,3,POSITIVE); 9 Using Arduino / Displays / Re: M2TKLIB Hello World with LiquidCrystal_I2C on: February 13, 2013, 11:26:47 pm I found this data on the connections of my display Quote ③PIN CONNECTIONS Pin Symbol Function) and tried to write a 'constructor' as follows (based on the library code I have copied at the bottom of this post Code: LiquidCrystal_I2C lcd ( 0x3F, 6, 5,4 , 7, 8, 9, 10, 15,POSITIVE); All I am getting at the moment is the backlight turns off and nothing is displayed (although there seems to be 2 lines of faint characters (this is a 4 line display) One thing I thought was odd is that it only has 4 data pins in the constructor, but the table above talks about 8. I did try pins 11 to 14 instead of 7 to 10, but it doesn't seem to make any difference. Code: /*! @method @abstract Class constructor. @discussion Initializes class variables and defines the I2C address of the LCD. The constructor does not initialize the LCD. @param lcd_Addr[in] I2C address of the IO expansion module. For I2CLCDextraIO, the address can be configured using the on board jumpers. @param En[in] LCD En (Enable) pin connected to the IO extender module @param Rw[in] LCD Rw (Read/write) pin connected to the IO extender module @param Rs[in] LCD Rs (Reset) pin connected to the IO extender module @param d4[in] LCD data 0 pin map on IO extender module @param d5[in] LCD data 1 pin map on IO extender module @param d6[in] LCD data 2 pin map on IO extender module @param d7[in] LCD data 3 pin map on IO extender module */ LiquidCrystal_I2C(uint8_t lcd_Addr, uint8_t En, uint8_t Rw, uint8_t Rs, uint8_t d4, uint8_t d5, uint8_t d6, uint8_t d7 ); // Constructor with backlight control LiquidCrystal_I2C(uint8_t lcd_Addr, uint8_t En, uint8_t Rw, uint8_t Rs, uint8_t d4, uint8_t d5, uint8_t d6, uint8_t d7, uint8_t backlighPin, t_backlighPol pol); 10 Using Arduino / Displays / Re: M2TKLIB Hello World with LiquidCrystal_I2C on: February 13, 2013, 10:54:05 pm Quote from: bperrybap on February 13, 2013, 07:01:58 pm There is varying i2c to hd44780 hardware out there. Some librarys are hard coded for a particular wiring/pin mapping and some libraries like fm's allow the pin mapping to be configured. So, In the case of fm's library the library is not the actual problem but rather the problem comes down to the user properly filling in the constructor which configures the pin mappings for the library. fm's library has a default output pin mapping for how the wires/traces are connected between the pcf8574 chip and the hd44780 interface. It is unwise not to fill in the full constructor and use the default mappings since different PCF8574 based boards wire up the connections differently. The default pin mapping in the library is for fm's lcd i/o expander board and many of the i2c based boards are not wired up the same way. When the pin mapping is wrong/incorrect, the library will compile just fine, but since it has been told the incorrect pin mappings, it will not be able to properly initialize and drive the lcd. Unfortunately, There is no way to have the library automatically detect the wiring and automatically set up the pin mappings. To ensure that i2c interface works with fm's library, it is best to always properly fill in the full constructor which includes all the pin mappings rather than only fill in the i2c address, which will cause the library use the pin mappings of fm's lcdio board, which is different from many of the other i2c to hd44780 boards out there. Plus, the default pin mappings do not include backlight control over the i2c interface since fm's lcd i/o board does not support this. If you fill in the full constructor, you can not only ensure that the pin mappings are correct,but also configure the library for how to control the backlight over the i2c interface. Well, thanks for filling me in on all that. I know nothing about constructors, and I really feel I have jumped in the deep end here. Still, it looks like there's hope of making it work, and I'm game to try. I need to find out where and how to set up the pin mappings, and what they really should be. I'll try looking the code and google, but if anyone can lead me in the right direction, that would help . 11 Using Arduino / Displays / Re: M2TKLIB Hello World with LiquidCrystal_I2C on: February 13, 2013, 05:17:48 pm My fault. I was sure they all worked when I first tried them, maybe I had an earlier version, or a modified one. When I googled for examples for my previous post, that one came up and I assumed it was the same. I am currently using the DFRobot one at the moment and it definitely works. I don't really understand how some can work, and others fail. Must be varying hardware out there too. Hopefully they will all eventually work as they iron out the bugs and add support for all the variations. 12 Using Arduino / Displays / Re: M2TKLIB Hello World with LiquidCrystal_I2C on: February 13, 2013, 01:58:53 am I got it to compile thanks. Unfortunately, that library just doesn't work with my display. Even their own HelloWorld_i2c example that comes with the Liquidcrystal library just produces a flashing backlight. The DFRobot library (and other libraries that I mentioned) does work with my display, so they've done something different. However they won't work with M2TKLIB because they lack things like LCD.h (I think). So I am stuck. 13 Using Arduino / Displays / Re: M2TKLIB Hello World with LiquidCrystal_I2C on: February 12, 2013, 08:16:02 pm Quote from: olikraus on February 12, 2013, 05:52:37 pm I added the New LiquidCrystal Library ( ) as output engine to M2tklib. A first beta release for the New LiquidCrystal Library is available for download: Oliver Thanks. I just downloaded it and am trying to understand how to make the Hello World example work with my I2C display. Firstly I had to include Wire.h to get it to compile. Then I tried to convert it to I2C by including the different library and changing how the lcd object was created (and changing the setup to 20x4 from 16x2) Code: ); } This is producing these errors Quote In file included from HelloWorld.pde:42: /home/myhome/sketchbook/libraries/M2tklib/utility/m2ghnlc.h:32: error: variable or field ‘m2_SetNewLiquidCrystal’ declared void /home/myhome/sketchbook/libraries/M2tklib/utility/m2ghnlc.h:32: error: ‘LCD’ was not declared in this scope /home/myhome/sketchbook/libraries/M2tklib/utility/m2ghnlc.h:32: error: ‘lc_ptr’ was not declared in this scope /home/myhome/sketchbook/libraries/M2tklib/utility/m2ghnlc.h:32: error: expected primary-expression before ‘cols’ /home/myhome/sketchbook/libraries/M2tklib/utility/m2ghnlc.h:32: error: expected primary-expression before ‘rows’ HelloWorld:45: error: no matching function for call to ‘Liquid ‘void setup()’: HelloWorld:51: error: ‘m2_SetNewLiquidCrystal’ was not declared in this scope So assuming I just don't understand I thought it was time to ask advice. 14 Using Arduino / Programming Questions / Re: Documentation for special/system variables/constants? on: February 10, 2013, 10:32:24 pm Quote from: PeterH on February 10, 2013, 10:23:27 pm. I see. So, if I read you correctly, the example I read on setting the timer to execute an interrupt is bypassing a normal way to do it. Obviously so did I . So I should look for a library the makes it easier to set up an interrupt routine then? Quote from: PeterH on February 10, 2013, 10:23:27 pm. Ah, so I will look up some references on C++ and GCC then. I guess it never occurred to me, being new to how it all works, that I could use native C++ stuff inside a sketch. 15 Using Arduino / Programming Questions / Re: Documentation for system variables/constants? on: February 10, 2013, 09:59:30 pm I am not looking for the documentation on the atmel chip. I am looking for the documentation of the language used to write sketches. It seems this language has special variables to access things - like the register example I gave, and also things like the PC system time (useful for setting RTC's at compile time). Another example would be the name of the routine it looks for when it encounters an interrupt. I just haven't found where they are documented. I looked here , but can't see it. This page has examples of special variables for many languages so you can see what I mean Pages: [ 1 ] 2 3 4 | SMF © 2013, Simple Machines Newsletter ©2014 Arduino
http://forum.arduino.cc/index.php?action=profile;u=160828;sa=showPosts
CC-MAIN-2014-10
refinedweb
2,418
61.46
This is a playground to test code. It runs a full Node.js environment and already has all of npm’s 1,000,000+ packages pre-installed, including gpe-portal-components-check with all npm packages installed. Try it out: require()any package directly from npm awaitany promise instead of using callbacks (example) This service is provided by RunKit and is not affiliated with npm, Inc or the package authors. This is a starter project for building a standalone Web Component using Stencil. Stencil is also great for building entire apps. For that, use the stencil-app-starter instead.. <script src=''></script>in the head of your index.html npm install my-component --save <script src='node_modules/my-component/dist/mycomponent.js'></script>in the head of your index.html npm install my-component --save import my-component;
https://npm.runkit.com/gpe-portal-components-check
CC-MAIN-2020-40
refinedweb
138
52.66
JavaBeans JavaBeans are the way to use Java objects from JSP pages in order to follow the MVC design pattern. The point of doing this is to implement something similar to the Pull methodology. For example: <jsp:useBean Examining the syntax of the above code, the first thing that pops up right away is the use of the scope attribute. How many HTML designers understand the programming concepts of scope? It is safe to suggest that a good portion of web designers barely understand the concept of how a CGI works. By stating this, we are not trying to slight people. Instead, we are simply pointing out that design and software engineering are distinct skill sets. You wouldn't expect a Java programmer to select a print and web safe color palette, would you? The common response to an argument like this is that the designers should simply ignore these tags and let others define and implement them. The problem with that is that you have now given them the power to accidentally wreck your entire application in such a way that it is very difficult to debug because a complex scope issue might not show up right away. The Java code: public class HelloBean { private String <jsp:setProperty </jsp:useBean> <HTML> <HEAD><TITLE>Hello</TITLE></HEAD> <BODY> <H1> Hello, <jsp:getProperty </H1> </BODY> </HTML> Above, we have a very simple example of using a bean in a page. Pass it some properties and then retrieve the results. This is the right way to do things when using JSP. However, if we look at an example of doing the same exact thing in Velocity, the extra amount of needless typing that one needs to perform to simply retrieve a property seems a just bit absurd. Of course there are always GUI based drag and drop tools to make typing a thing of the past. Really. There are several commercial solutions available today which provide a nice drag and drop view for doing development with JSP and Struts. However many of these tools are still first generational tools. They typicially only address parts of the problem and require digging down into the nitty gritty stuff when things become difficult or even impossible to do with the GUI (anyone remember a product called Tango?). Often these tools also produce code that is not optimized for heavily hit sites and getting an existing application to scale sometimes requires a complete rewrite. Again, this is not our decision, it is yours. Another item to note here is that these are costly (>$1000/seat) development tools. In this .bomb economy, who really has the money to spend on these tools? The Java code: context.put ("hello", new HelloBean()); The Velocity code: $hello.setName("*") <HTML> <HEAD><TITLE>Hello</TITLE></HEAD> <BODY> <H1> Hello, $hello.Name </H1> </BODY> </HTML> The example shows the creation of the HelloBean object and then placing it into the Context. Then, during runtime execution of the template, that object is available as a $variable which uses the JavaBean specification to do introspection on the object. For example, Velocity uses Bean style introspection to permit the method call to be shortened from $hello.getName() to simply typing what is shown above. When Velocity is combined with Turbine, the HelloBean object can be added into the Context as a configuration option or it can be added at any point of the processing. This is what provides the "scope" of the object in the Context. Another "gotcha" with using JavaBeans in JSP is again quoted from Jason's book: One thing to watch out for: On some servers (including Tomcat 3.2) if you have a bean with a scope of "session" or "application" and you change the bean class implementation, you may get a ClassCastException on a later request. This exception occurs because the generated servlet code has to do a cast on the bean instance as it's retrieved from the session or application, and the old bean type stored in the session or application doesn't match the new bean type expected. The simplest solution is to restart the server. You make the decision. YmtdErrorHandling <- Previous | Next -> YmtdSampleApplication
https://wiki.apache.org/velocity/YmtdJavaBeans?action=diff
CC-MAIN-2018-05
refinedweb
696
61.46
Nowadays JSON is a very common choice for data interchange on the web. A lot of REST APIs use JSON-encoded body in their requests and responses. The default choice for dealing with JSON in Haskell is the aeson package. So it’s not surprising that most nontrivial Haskell projects depend on aeson, directly or indirectly. Typeable heavily relies on aeson too. As a core package everyone depends on, aeson is relatively well optimised. Yet there is always room for improvement. Given that even slight speedup in a core package will have a big impact on the whole community, we at Typeable decided to invest some time into aeson performance. In this post we’ll describe one trick we applied to speedup the JSON parser in this pull request. One of the most performance-critical parts of any JSON parser is string unescaping. The specification requires not printable characters, the double quote and the backslash to be escaped. E.g. the string a"b is represented as "a\"b". Also, an arbitrary character can be represented using the Unicode code point, e.g. \u045E represents the CYRILLIC SMALL LETTER SHORT U , namely ў. When parsing JSON, we have to unescape strings back. Doing it naïvely may introduce a performance bottleneck because unescaping is stateful: when scanning the string we need to remember whether the previous character was the backslash. But most of the strings in a typical JSON are not escaped, so the naïve implementation will make the typical case unnecessary slow. Here the slow path trick comes into play. Let’s describe the trick in a more general setting. Suppose we have a complex and slow, but correct algorithm. And we have a simpler and faster algorithm, which covers only the typical case, but isn’t applicable for the general case. The trick is to pretend that only the typical case exists and apply the fast algorithm (the common path), but fallback to the slower one when necessary (the slow path). In the case of string unescaping in a JSON parser the common path is when unescaping is not needed, and the slow path actually unescapes the string. Separating the common and the slow paths has two benefits. Since we apply a faster algorithm for the typical case, the whole performance improves. But there is also an interesting side effect. Often the common and the slow paths have a different set of trade-offs. If we mix both in one algorithm, then we have to sacrifice the general case performance in order to improve the typical case performance. As a result, we often get an implementation that is optimal neither for the general, nor for the typical case. Separating them, on the other hand, allows us to optimize the paths independently, improving performance even for the general case! Let’s see how to apply the trick to the JSON parser in aeson. The parser is implemented using the attoparsec package. Parser for strings looks like the following: -- | Parse a string without a leading quote. jstring_ :: Parser Text jstring_ = do (s, S _ escaped) <- A.runScanner startState go <* A.anyWord8 -- We escape only if there are -- non-ascii (over 7bit) characters or backslash present. if isTrue# escaped then case unescapeText s of Right r -> return r Left err -> fail $ show err else return (TE.decodeUtf8 s) where startState = S 0# 0# go (S skip escaped) (W8# c) | isTrue# skip = Just (S 0# escaped') | isTrue# (w ==# 34#) = Nothing -- double quote | otherwise = Just (S skip' escaped') where w = word2Int# c skip' = w ==# 92# -- backslash escaped' = escaped `orI#` (w `andI#` 0x80# ==# 0x80#) -- c >= 0x80 `orI#` skip' data S = S Int# Int# (You can find the full code here) The idea here is simple: we scan the input until we reach the closing quote, then we unescape the string. The skip flag in the go helper makes sure we won’t stop on an escaped double quote character. The code is manually optimized, i.e. the go helper manipulates unboxed values. Also, the code already tries to optimize for the common case: it applies the unescapeText function only when unescaping is needed. That’s why we need the escaped flag: we don’t need to unescape the string unless we saw the backslash or any non-ascii character in it. There are two problems in the code above. We use the stateful scanner even for the typical case. Maintaining the skip and escaped flags takes CPU circles and slows down the parser even for strings that are not escaped. Let’s apply the slow path trick described above and pretend that we never need to unescape strings. In that case, the algorithm is quite simple: we just take the input until we reach the closing double quote: jstring_ = do s <- P.takeWhile (\w -> w /= DOUBLE_QUOTE) return (TE.decodeUtf8 s) This code is much faster, though it fails when the string is escaped. We need a way to detect that we need to fallback. In this particular case it’s easy: we stop not only on the double quote, but also on the backslash and any non-ascii character. If we stopped on the double quote, then we are OK, otherwise we need to fallback: jstring_ = do s <- A.takeWhile (\w -> w /= DOUBLE_QUOTE && w /= BACKSLASH && not (testBit w 7)) w <- A.peekWord8 case w of Just DOUBLE_QUOTE -> A.anyWord8 $> TE.decodeUtf8 s _ -> slowPath s -- here we fallback to the general case Here the slowPath is basically a call to the original jstring_ function, modified to take the already consumed part of the input into account. Applying this optimization improves parser performance by 15%-45% depending on how many escaped strings the input contains. As mentioned above, separating the slow path may open a possibility to improve the slow path performance. It’s true in our case too. Note that we don’t need the escaped flag anymore because we already know that the string is escaped. It’s the slow path after all. Removing the flag eliminates the need for manual unboxing, now the idiomatic implementation is as fast as the hand-optimized one: slowPath s' = do s <- A.scan startState go <* A.anyWord8 case unescapeText (B.append s' s) of Right r -> return r Left err -> fail (show err) where startState = False go a c | a = Just False | c == DOUBLE_QUOTE = Nothing | otherwise = let a' = c == BACKSLASH in Just a' I didn’t noticed any speedup after this change, but at least it simplified the code and improved maintainability. The slow path trick is not a silver bullet and can be applied only when certain conditions are met. Obviously, there should exist a typical case, which is common enough to make the optimization worthwhile. But also there should be a cheap way to detect that we need to fallback to the general algorithm, so that the check won’t slow down the common path too much. Of course, the described slow path optimization trick is not new. I don’t remember where exactly I learned it from, but I’m sure it’s widely known. Yet people often overlook this simple but powerful optimization, so I think it’s beneficial to draw more attention to it. I hope you’ll spot a possibility to apply the trick the next time you’ll work on performance optimization. The aeson package is already optimized for performance, and it’s a challenge to find a space for improvement here. So I think 15%-45% speedup is something to be proud of. But we are not going to stop here. We are already working on another optimization, and our preliminary measurements are quite promising. So stay tuned!.
https://typeable.io/blog/2020-02-24-performance_slow_path.html
CC-MAIN-2022-05
refinedweb
1,279
63.7
[ ] Advertising Andy Pook commented on LUCENENET-469: ------------------------------------- Yeah, well, I thought, that's looks like it might be isolated, small. Ha! I'll poke at it some more, see what turns up. But I may just punt and look for something else :) I'm not sure I'm on the same wave length just yet... "how does that apply to TermsEnum? Disposing a superclass is not quite the same thing" But the enumerator is not a super class. It's a "child" of the TermsEnum (or whatever) created for a purpose. Disposing the child shouldn't dispose the "parent". I'm fairly sure that IEnumerator has Dispose so that things like listing files (managed resource) can be cleaned up. But still, it's about cleaning up resources used by the enumerator _not_ by the parent collection type. If you want to Dispose the collection type then that's a separate operation. There may be other behaviors available, appropriate on the collection type _after_ you've looped some aspect of it. Right ?? Also a bit confused around the "extras" topic. I don't see the problem. We'd get all the linq bits for free. You're right, they are generally defined on {{IEnumerable<T>}} But that's fine. In your example of {{TermsEnum}} being {{IEnumerable<BytesRef>}} then regardless of how the enumerator is factored you can still do {{foreach(var item in termsEnum.Where(b=>b.Length>100)){}}} > Convert Java Iterator classes to implement IEnumerable<T> > --------------------------------------------------------- > > Key: LUCENENET-469 > URL: > Project: Lucene.Net > Issue Type: Sub-task > Components: Lucene.Net Contrib, Lucene.Net Core > Affects Versions: Lucene.Net 2.9.4, Lucene.Net 2.9.4g, Lucene.Net 3.0.3, > Lucene.Net 4.8.0 > Environment: all > Reporter: Christopher Currens > Fix For: Lucene.Net 4.8.0 > > > The Iterator pattern in Java is equivalent to IEnumerable in .NET. Classes > that were directly ported in Java using the Iterator pattern, cannot be used > with Linq or foreach blocks in .NET. > {{Next()}} would be equivalent to .NET's {{MoveNext()}}, and in the below > case, {{Term()}} would be as .NET's {{Current}} property. In cases as below, > it will require {{TermEnum}} to become an abstract class with {{Term}} and > {{DocFreq}} properties, which would be returned from another class or method > that implemented {{IEnumerable<TermEnum>}}. > {noformat} > public abstract class TermEnum : IDisposable > { > public abstract bool Next(); > public abstract Term Term(); > public abstract int DocFreq(); > public abstract void Close(); > public abstract void Dispose(); > } > {noformat} > would instead look something like: > {noformat} > public class TermFreq > { > public abstract Term { get; } > public abstract int { get; } > } > public abstract class TermEnum : IEnumerable<TermFreq>, IDisposable > { > // ... > } > {noformat} > Keep in mind that it is important that if the class being converted > implements {{IDisposable}}, the class that is enumerating the terms (in this > case {{TermEnum}}) should inherit from both {{IEnumerable<T>}} *and* > {{IDisposable}}. This won't be any change to the user, as the compiler > automatically calls {{IDisposable}} when used in a {{foreach}} loop. -- This message was sent by Atlassian JIRA (v6.4.14#64029)
https://www.mail-archive.com/dev@lucenenet.apache.org/msg03723.html
CC-MAIN-2017-34
refinedweb
500
65.32
In this article we will learn to pass one class as a parameter of another class. In broad sense ,we are trying to create parameterized class where we will pass another class as parameter and we will use the property of the parameter class. This article demands little knowledge of parameterized class. If you are completely new in this concept then below paragraph is for quick understanding of same. What is parameterized class? In case of function we use and pass parameter to pass value. In C# we can pass parameter to a class and the class which takes parameter is called parameterized class or generic class. Try to understand below code. class TestClass<T> { //Create object of parameterized class. T obj; } This is small example of parameterized class and in place of T we can pass any object, event it may be another class. In below example we will try to implement one complete example of parameterized class. Have a look on below code. In this above example we created two classes. The Second class is our main class which we are passing as a parameter of First class. Here is code of Main() function. Second s = new Second(); s.name = "sourav"; s.surname = "kayal"; First<Second> objf = new First<Second>(); At first we are creating object of “Second” class and then setting values to properties, after that we are creating object of “First” class by passing “Second” class as object. But the problem is that we are not able to access the property of Second class within first class. In above example After T( Type) we are giving “.” But IDE is not showing any property. If we want to use property of parameter class then we have to use “where” keyword. Try to understand below code. using System;using System.Collections.Generic;using System.Linq;using System.Text;using System.Threading.Tasks;using System.Diagnostics;using System.Threading;namespace ConsoleApp{ class First<T> where T:Second { public First(T tmp) { Console.WriteLine("Name is : " + tmp.name); Console.WriteLine("Surname is : " + tmp.surname); } } class Second { public string name { get; set; } public string surname { get; set; } } class Program { static void Main(string[] args) { Second s = new Second(); s.name = "sourav"; s.surname = "kayal"; First<Second> objf = new First<Second>(s); Console.ReadLine(); } }} Here is sample output. In this quick article we have learned how to pass class as a parameter of another class. Hope you have learned the concept. Latest Articles Latest Articles from Sourav.Kayal Login to post response
http://www.dotnetfunda.com/articles/show/2641/pass-csharp-class-as-a-parameter-of-another-class
CC-MAIN-2019-43
refinedweb
420
60.41
How you can build a Serverless API using GraphQL .Net Core, C# and VS Code Chris Noring Updated on ・13 min read Follow me on Twitter, happy to take your suggestions on topics or improvements /Chris This article takes you through how to build a full CRUD API using GraphQL. We end up hosting it in a Serverless function and also show how you can do external HTTP calls to build your API. Whether the data lives here or somewhere else doesn't matter. GraphQL is the front that your users will talk to. TLDR; This article might be a bit long but it does teach a lot on GraphQL, queries, mutations. It also teaches you how to create a serverless function all in the context of .Net Core, C# and VS Code In this article we will: - Learn how to build a full CRUD GraphQL including queries and mutations - Create a Serverless function and Host our GraphQL API in a function - Show how you can make an external HTTP call and make that part of our GraphQL API Resources Getting started with GraphQL in .Net This is the first article I've written on the topic of GraphQL with .Net. I recommend you have a look to get a feeling for the basics of GraphQL. - Starting with Azure Functions Covers creating a Serverless function and deploying to the Cloud using VS Code Starting with Serverless in .Net .Net C# Developer reference to Azure Functions Creating a Serverless API in C# and .Net Article I wrote covering how to create a CRUD API with Serverless, .Net, C#, and VS Code Create your first Serverless function LEARN module for creating an Azure Function Free Azure account To deploy Serverless Azure Functions you will need a free Azure account) Create a Serverless function app The first thing we will do is to create a Serverless function. But what is Serverless and why might it be useful in the context of GraphQL? That's really two questions but let's try to answer them in order. Why Serverless? Serverless is not about no servers but more about the server is NOT in your basement, anymore. In short, the servers have moved to the Cloud. That's not all, however. Serverless have come to mean other things namely that everything is set up for you. That means you don't need to think about what OS your code runs on or what Web Server runs it, everything is managed. There is more. There is always more, isn't it? What now a fancy sticker? It's about cost. Serverless usually means it's cheap. Why is it cheap? Well, Serverless code is meant to be run seldom and you are only billed for the time it executes Ok, how do YOU make money? Well, the function isn't always there. Whatever resources are needed are allocated when someone/something queries your function. Doesn't that make it a bit slow like there could be some initial wait when it's scrambling to be created and then respond the query? You are right. This is something called a cold start. This can be avoided however by either regularly polling our function or using a premium offering that shortens the cold start. Ok so it's I can either have 100% availability or cheap, pick one :) I guess so. Why Serverless with GraphQL? Ok, we are a bit smarter about the whats and whys of Serverless so why add GraphQL to the mix? Well, GraphQL has the ability to stitch together data from different APIs so it could act as an API that aggregates data from different sources. So in a sense, it works well in the context of Serverless cause it's not like you need to have any data stored in the function if it's someone else's API. Ok sounds good, I guess there are no other good reasons the rest is just about the hype that's GraphQL and Serverless right? ;) ... Prerequisits To create a Serverless function we first need a function app to put it in. To makes this as easy as possible we will be using VS Code and an Azure Function extension. So our prerequisites are: - Node.js - Visual Studio Code - Azure Functions extension We can download Node.js from the following page: You can find Visual Studio code here: As for the extension we need. Search for an extension called Azure Functions. It should look like the below: Scaffold Ok, now we should be all setup and ready to create our first Serverless function. As mentioned before, the function needs to live within something called a function app. So we need to create the function app first. Hit CMD+SHIFT+P or choose View/Command Palette to bring up the command palette: Now we need to type a command that will help us create a Serverless function app. It's called Azure Functions: Create New Project Then it's going to ask you for what directory to create the app. The directory you are standing in is the default chosen one, select that. Next thing it asks about is the language, select C#. Next up it's asking you about the projects first function and how it should be triggered. Select HttpTrigger. Now you need to provide a function name. Call it Graphql. Next question is a namespace. You could really choose anything here but let's call it Function just for the sake of it. Lastly, it asks about Access Right. There are different options here Anonymous, Function, Admin. We select Anonymous as we want to make our function publically available. The other options mean we would need to provide some kind of key when calling the function. VS Code should now scaffold all the needed files. It should also ask you to restore all dependencies, to download the libraries. You could also run dotnet restore if you should miss clicking on this dialog. Your project structure should now look like this: Test it out Let's ensure that we can run our newly scaffolded function. First, you should have been prompted if you want to add the necessary to be able to debug. You should answer YES here. This will generate a directory .vscode looking like the below: The tasks.json contains tasks needed to build, clean and run your project. These will assist you with the step we are about to do next Debugging. To Debug we choose Debug/Start Debugging from the menu. It should compile the code and once done it should look like this: It's telling us to go to Let's spin up a browser: Seems to work alright. :) It's hitting our function Graphql. Speaking of which, let's have a quick look at the sample code we've been given when we scaffolded the Serverless app: // Graphql.cs"); } } } Let's not spend to much time understanding everything right now but suffice to say it does its job and is able to work with query parameters: string name = req.Query["name"]; and the Body, in case you POST a request: string requestBody = await new StreamReader(req.Body).ReadToEndAsync(); dynamic data = JsonConvert.DeserializeObject(requestBody); name = name ?? data?.name; Next, let's talk GraphQL. Adding GraphQL to Serverless Ok, we established that we wanted to add GraphQL to our Serverless function. First, let's build a GraphQL API. Any GraphQL will contain the same moving parts: - A Schema, this will define what we can query for and what custom data types we have - A Resolver, a collection of functions that is able to respond to a request and ends up delivering a response Adding a GraphQL schema This will be authored in something called GQL or GraphQL Query Language. Our schema will look like the following: } Now that's a lot. Let's explain what we are looking at. Query Anything of type Query or Mutation is something we can ask for in our query to our API. The semantic meaning of Query is that you want to fetch data. In this case, we want to fetch a list of jedis using a query syntax like this: { jedis { name side } } This is equivalent to writing the following in SQL: SELECT name, side FROM jedis; The other thing we do is supporting a query with a parameter, namely jedi(id: ID): Jedi. We call it like so: { jedi(id: 1) { name } } This is equivalent to writing the following in SQL: SELECT name FROM jedi WHERE id=1; Custom types Everything queryable was defined under type Query. Everything else except for type Mutation are custom types that we define. For example: type Jedi { id: ID name: String, side: String } Mutation This semantically means we will try to change the data. Looking at the operations we support: addJedi(input: JediInput): Jedi updateJedi(input: JediInput ): Jedi removeJedi(id: ID): String We can see that we support adding, updating and removal. To call a mutation we need to type like so: mutation test { addJedi(input: { name: "JarJar", side: "Dark" }) { name } } Input, complex input for our Mutation Note the input property input in addJedi(). If we look at the schema we can see that it's of type JediInput which is defined like this: input JediInput { name: String side: String id: ID } The reason we are using the keyword input instead of type is that this is a special case. Special how, you wonder? Well, there are two types of input parameters to a mutation: - Scalars, this is a String, ID, Boolean, etc, also called primitives - Inputs, this is nothing more than a complex data type with many properties on it, example JediInput so it's definitely possible to define a mutation looking like this: type Mutation { addTodo(todo: String!): String } So the million-dollar question is why can't I just use a custom type like Jedi as an input parameter type to my mutation? The honest answer is I don't know. Just remember this: If you need an input parameter that is more complex than a scalar, then you need to define it like so: input MyInputType { // my columns } Add NuGet package Ok next step is to set up this schema properly in code. For that we will create a file Server.cs and also install the GraphQL package like so: dotnet add package GraphQL Create a Schema Now add the following code to Server.cs, like so: using GraphQL; using GraphQL.Types; using Newtonsoft.Json; using System.Threading.Tasks; namespace Function { public class Server { private ISchema schema { get; set; } public Server() { this.schema = Schema.For(@" } ", _ => { _.Types.Include<Query>(); _.Types.Include<Mutation>(); }); } public async Task<string> QueryAsync(string query) { var result = await new DocumentExecuter().ExecuteAsync(_ => { _.Schema = schema; _.Query = query; }); if(result.Errors != null) { return result.Errors[0].Message; } else { return JsonConvert.SerializeObject(result.Data); } } } } In the constructor above we set up the Schema by calling Schema.For() with a string representing our schema expressed as GQL. The second argument of that won't compile though, namely this part: _ => { _.Types.Include<Query>(); _.Types.Include<Mutation>(); } Adding resolvers The reason is it won't compile is that Query and Mutation don't exist yet. They are simply resolver classes answering to query and mutation requests. Let's make it compile by creating first Db.cs and a file Query.cs Adding in-memory database // Db.cs using System.Collections.Generic; using System.Linq; namespace Function { public class StarWarsDB { private static List<Jedi> jedis = new List<Jedi>() { new Jedi(){ Id = 1, Name ="Luke", Side="Light"}, new Jedi(){ Id = 2, Name ="Yoda", Side="Light"}, new Jedi(){ Id = 3, Name ="Darth Vader", Side="Dark"} }; public static IEnumerable<Jedi> GetJedis() { return jedis; } public static Jedi AddJedi(Jedi jedi) { jedi.Id = jedis.Count + 1; jedis.Add(jedi); return jedi; } public static Jedi UpdateJedi(Jedi jedi) { var toUpdate = jedis.SingleOrDefault(j => j.Id == jedi.Id); toUpdate.Name = jedi.Name; toUpdate.Side = jedi.Side; return toUpdate; } public static string RemoveJedi(int id) { var toRemove = jedis.SingleOrDefault(j => j.Id == id); jedis.Remove(toRemove); return "success"; } } public class Jedi { public int Id { get; set; } public string Name { get; set; } public string Side { get; set; } } } Db.cs is nothing more than a simple in-memory database. Adding a resolver class to handle all Queries Next, let's create Query.cs: // Query.cs using System.Collections.Generic; using GraphQL; using System.Linq; namespace Function { public class Query { [GraphQLMetadata("jedis")] public IEnumerable<Jedi> GetJedis() { return StarWarsDB.GetJedis(); } [GraphQLMetadata("jedi")] public Jedi GetJedi(int id) { return StarWarsDB.GetJedis().SingleOrDefault(j => j.Id == id); } [GraphQLMetadata("hello")] public string GetHello() { return "World"; } } } What's interesting about the above is how we map something in our GraphQL schema to a resolver function. For that we use the decorator GraphQLMetadata like so: [GraphQLMetadata("jedis")] public IEnumerable<Jedi> GetJedis() { return StarWarsDB.GetJedis(); } The above tells us that if the user queries for jedis then the function GetJedis() will respond. Dealing with parameters is almost as easy. It's the same decorator but we just add input parameter like so: [GraphQLMetadata("jedi")] public Jedi GetJedi(int id) { return StarWarsDB.GetJedis().SingleOrDefault(j => j.Id == id); } Adding a resolver class to handle all Mutation requests We are almost there but we need the class Mutation.cs and let's add the following content to it: // Mutation.cs using GraphQL; namespace Function { public class Mutation { [GraphQLMetadata("addJedi")] public Jedi AddJedi(Jedi input) { return StarWarsDB.AddJedi(input); } [GraphQLMetadata("updateJedi")] public Jedi UpdateJedi(Jedi input) { return StarWarsDB.AddJedi(input); } [GraphQLMetadata("removeJedi")] public string AddJedi(int id) { return StarWarsDB.RemoveJedi(id); } } } As you can see it looks very similar to Query.cs, even with parameter handling. Updating our Serverless function There's just one piece left of the puzzle, namely our serverless function. It needs to be altered so we can support the user to invoke our function like so: url?query={ jedis } { name } Change the code in Graphql.cs to the following:) { var server = new Server(); string query = req.Query["query"]; // string query = "mutation test { addJedi(input: { name: \"JarJar\", side: \"Dark\" }) { name } }"; var json = await server.QueryAsync(query); return new OkObjectResult(json); } } } Test it all out To test it we just start debugging by choosing Debug/Start Debugging from the menu and change the URL in the browser to:{ jedis { name } } Let's see what the browser says: Yep, we did it. Serverless function with a GraphQL API. Bonus - calling other endpoints Now. One of the great things about GraphQL is that it allows us to call other APIs and thereby GraphQL acts as an aggregation layer. We can easily achieve that by the following steps: - Do the HTTP request, this should do the request to our external API - Add resolver method for our external call - Update our schema with the new type - Test it out HTTP Request We can easily do HTTP request in .Net with HttpClient so let's create a class Fetch.cs like so: // Fetch.cs using System.Net.Http; using System.Threading.Tasks; namespace Function { public class Fetch { private static string BaseUrl = ""; public static async Task<string> ByUrl(string url) { using (var client = new HttpClient()) { var json = await client.GetStringAsync(string.Format("{0}/{1}", BaseUrl, url)); return json; } } } } Add resolver method Now open up Query.cs and add the following method to the Query class: [GraphQLMetadata("planets")] public async Task<List<Planet>> GetPlanet() { var planets = await Fetch.ByUrl("planets/"); var result = JsonConvert.DeserializeObject<Result<List<Planet>>>(planets); return result.results; } Additionally, we should be installing a new NuGet package: dotnet add package Newtonsoft.Json This is needed so we can convert the JSON response we get into a Poco. We should also add types Planet and Result to Query.cs, but outside the class definition: public class Result<T> { public int count { get; set; } public T results { get; set; } } public class Planet { public string name { get; set; } } Update our schema with the new type We have the schema left to update before we can try it out. So let's open up Schema.cs and make sure it now looks like this: this.schema = Schema.For(@" type Planet { name: String } planets: [Planet] } ", _ => { _.Types.Include<Query>(); _.Types.Include<Mutation>(); }); Above we've added the type Planet: type Planet { name: String } We've also added planets to Query: planets: [Planet] Testing it out That should be it, let's test it. Debug/Start Debugging:{ planets { name } } There we have it. :) We've managed to do an external call to an API and make that part of our GraphQL API. As you can see we can intermix data from different sources easily. Summary We've managed to create a GraphQL that supports the full CRUD. That is we support both querying for data but also creating, update and delete. Additionally, we've taken our first steps with Serverless functions. Finally, we added our GraphQL API to be served by our Serverless function. As an added bonus we also showed how we could take to external APIs and make them part of our API. That's really powerful stuff. I hope you enjoyed this article. In out next part we will take a look at how to add GraphQL to a Web API project in .Net Core. Hi, Nice article, thanks 😀 I played with GraphQL a bit and what really put me off is the schema being defined in a string, so no strong typing, no coherence between the schema and the code etc It's kind of ok for jedi and Todo examples but a full blown API will become very difficult to manage. Does anyone know of an implementation that doesn't rely on schema in a string? well I think the way they solve that is by having extensions like this one marketplace.visualstudio.com/items... so you at least have tooling support. I mean it is strongly typed. You get quite good indications that you are doing it wrong if you start typing in GraphIQL, the UI environment Wow, another fantastic, in-depth article Chris ! It's always a pleasure to read your work :) Thank you Mark :) much appreciated Great article. I still think we have a long way to go with developer friendliness around these types of APIs. The number of moving parts and concepts that need to be grasped for basic crud borders on intellectual masturbation IMHO. I’d like to see someone build a framework that generates or abstracts all of this to require way less code. Great job breaking it down and showing us the state of these technologies! yea I agree. There definitely exists a lot of great generators for generating code from schemas. That will have to be a separate article :) Excellent work @softchris , it's always a glad to read your posts!!! Very very good! Follow +1 <3 thank you, Jeffrey. Really appreciate that :) Thanks for the article. I am excited about GraphQL. We are currently using OData heavily. I hope GraphQL can be a good alternative or at least be a complementary for OData.
https://practicaldev-herokuapp-com.global.ssl.fastly.net/azure/how-you-can-build-a-serverless-api-using-graphql-net-core-c-and-vs-code-g5h
CC-MAIN-2020-10
refinedweb
3,182
65.32
Saturday, August 27, 2016¶ Today we (my family and I) will fly back home. I wanted a paper copy of our boarding passes, but without the lower part which contains only publicity. I installed PyPDF2 and wrote the following script: from __future__ import print_function from PyPDF2 import PdfFileWriter, PdfFileReader from argh import dispatch_command, arg, CommandError @dispatch_command @arg('filenames', help="The pdf files to crop.") @arg('--bottom', default=None, help='Height of bottom area to crop. Default is half of page.') def main(bottom=None, *filenames): output = PdfFileWriter() for filename in filenames: pdfin = PdfFileReader(open(filename, "rb")) # print how many pages pdfin has: print("{} has {} pages.".format(filename, pdfin.getNumPages())) for p in pdfin.pages: mb = p.mediaBox # print(mb) b = bottom or (mb.getUpperRight_y() / 2) # p.mediaBox.upperRight = ( mb.lowerRight = (mb.getLowerRight_x(), b) mb.lowerLeft = (mb.getLowerLeft_x(), b) output.addPage(p) output.write(file("output.pdf", "wb")) The script produced appearently what I want, but in the printout the cropped area then showed up again. IOW it needs more work, but I won’t do this now since I have my paper copies for this time.
http://luc.lino-framework.org/blog/2016/0827.html
CC-MAIN-2018-39
refinedweb
186
51.65
How to set namespace prefix for JAX-WS generated WSDL file? Hello, I have developped a small web service in contract first mode. So I have created my class with @WebService annotation. My webservice has one operation which takes a "Credentials" bean as input. It is defined in the same Java package. On my bean I use @XmlRootElement and @XmlElement to specify the namespace, such as namespace="". In the WSDL, I would like to be able to explicitly specify the prefix for this namespace but I am unable to do that. I tried by adding a package-info.java file with: @XmlSchema(namespace = "", elementFormDefault = javax.xml.bind.annotation.XmlNsForm.QUALIFIED, xmlns = { @XmlNs(prefix = "t", namespaceURI = "")}) But nothing is changed in the WSDL. Does anyone know how to do that ? Thanks a lot. Jena Hi Redema, Thank you for your answer. But my point is not to set the namespace but to set the prefix namespace. Regards, JB Hi Jena, sorry i have misread your post. I did some search and tests myself but can not get it working as well. Strange: this is the way it should be. On the other hand: why you want a specific prefix where the default/generated is perfectly fine. Its valid xml. The client will always understand the default/generated prefix. Martin. I fully agree that specifying a prefix is not really usefull as the default/generated one is perfectly fine. But my final customer is complaining of the namespaces because he is using SoapUI to create some SOAP requests and with the default namespace it is cosmeticly not really readable. Jena Hi Jena, looks like you only need the specify the "better readable" namespace. No need for the prefix. Maybe the package-info.java is ignored because we use the @WebService() annotation. Sorry I can not be more helpful. Martin. Hi Jena, in the @WebService annotation you can specify the wsdl namespace by adding the attribute "targetNamespace=" like in @WebService(serviceName = "TestService", targetNamespace="") Hope this helps. Regards, Martin.
https://www.java.net/forum/topic/glassfish/metro-and-jaxb/how-set-namespace-prefix-jax-ws-generated-wsdl-file
CC-MAIN-2015-11
refinedweb
335
59.9
Provided by: manpages-dev_4.16-1_all NAME clone, __clone2 - create a child process SYNOPSIS /* Prototype for the glibc wrapper function */ #define _GNU_SOURCE #include <sched.h> int clone(int (*fn)(void *), void *child_stack, int flags, void *arg, ... /* pid_t *ptid, void *newtls, pid_t *ctid */ ); /* For the prototype of the raw system call, see NOTES */ virtual address space, the table of file descriptors, and the table of signal handlers. (Note that on this manual page, "calling process" normally corresponds to "parent process". But see the description of CLONE_PARENT below.) One use of clone() is to implement threads: multiple flows of control in a program that run concurrently in a shared address space. When the child process is created with clone(),: CLONE_CHILD_CLEARTID (since Linux 2.5.49) Clear (zero). The store operation completes before clone() returns control to user space.. This flag is intended for the implementation of containers. (veth(4)) device. Only a privileged process (CAP_SYS_ADMIN) can employ CLONE_NEWNS. It is not permitted to specify both CLONE_NEWNS and CLONE_FS in the same clone() call. For further information on mount namespaces, see namespaces(7) and mount_namespaces(7). namespaces(7) and.) The store operation completes before clone(). Since then, the kernel silently ignores this bit if it is specified in flags. CLONE_PTRACE (since Linux 2.2) If CLONE_PTRACE is specified, and the calling process is being traced, then trace the child also (see ptrace(2)). CLONE_SETTLS (since Linux 2.5.32) The TLS (Thread Local Storage) descriptor is set to newtls. The interpretation of newtls and the resulting effect is architecture dependent. On x86, newt). NOTES Note that the glibc clone() wrapper function makes some changes in the memory pointed to by child. Another difference for the raw clone() system call is that the child_stack argument may be NULL, in which *child_stack, int *ptid, int *ctid, unsigned long newtls); On x86-32, and several other common architectures (including score, ARM, ARM 64, PA-RISC, arc, Power PC, xtensa, and MIPS), the order of the last two arguments is reversed: long clone(unsigned long flags, void *child_stack, int *ptid, unsigned long newtls, int *ctid); On the cris and s390 architectures, the order of the first two arguments is reversed: long clone(void *child_stack, unsigned long flags, int *ptid, int *ctid, unsigned long newtls); On the microblaze architecture, an additional argument is supplied: long clone(unsigned long flags, void *child_stack, int stack_size, /* Size of stack */ int *ptid, int *ctid, unsigned long newt the glibc clone() wrapper function when fn or child_stack is specified as NULL.. EINVAL child_stack is not aligned to a suitable boundary for this architecture. For example, on aarch64, child_stack must be a multiple of 16. ENOMEM Cannot allocate sufficient memory to allocate a task structure for the child, or to copy those parts of the caller's context that need to be copied. flags, and the limit on the number of nested user namespaces would be exceeded. See the discussion of the ENOSPC error above. CONFORMING TO clone() call to clone().). For a while there was CLONE_DETACHED (introduced in 2.5.32): parent wants no child-exit signal. In Linux 2.6.2, the need to give this flag together with CLONE_THREAD disappeared. This flag is still defined, but has no effect..16 of the Linux man-pages project. A description of the project, information about reporting bugs, and the latest version of this page, can be found at.
http://manpages.ubuntu.com/manpages/cosmic/man2/clone.2.html
CC-MAIN-2019-30
refinedweb
565
65.22
Not every library will work on Windows Phone 7 Not every library will work on Windows Phone 7 Join the DZone community and get the full member experience.Join For Free As obvious as this might seem, there are still lots of questions online from people asking whether it is possible to use general .NET-targeted libraries in Windows Phone 7. The common misconception here is that if a library is built on top of the .NET Framework, it should work in any .NET project regardless of the platform. As good as this might sound, it is not entirely true. Although the fundamentals of the framework still apply, like the base types, a lot of functionality will be different (e.g. missing methods or classes), and there is a reason for that. Developers at Microsoft did not remove functionality for the sake of minimizing the set of available tools. The main reasons for this are: - Performance - some classes and the associated calls have a significant impact on system resource consumption levels, therefore using them on a mobile platform would not be a good idea because mobile devices have limited capabilities. For a good example, take a look at methods asociated with HttpWebRequest, WebClient or with a passed service instance. When it comes to data consumption, those are all asynchronous and synchronous methods are either removed completely or have their async implementation. - Security - specifically, this applies to cross-domain and device security. For example, you cannot access the filesystem on the phone the same way you did when you were developing desktop or even web applications. This is introduced to protect other applications (and the OS itself) from tampering and piracy. Another example could be the fact that you cannot directly load a XML file in XDocument right from the web - only local paths are acceptable. That being said, since Windows Phone 7 is lacking specific "big machine" functionality like databases (SQL and non-SQL), you cannot access the System.Data namespace and all external database access should be done via web services. This applies to lots of other namespaces as well (e.g. System.Xml, whcih in fact is much more limited on WP7). You should also remember that if a library is not listed in the References dialog by default (e.g. System.Json) this doesn't yet mean that you cannot use it on the phone. Some additional libraries are located in the Silverlight SDK folder on your local machine and you should check that location (I wrote about this possibility here). So what exactly can you do to make sure that a specific library works in a Windows Phone 7 application? The best option is if the library was originally built as a Windows Phone 7 library. Other than that, an alternative would be building a Silverlight library. However, remember that even if you are able to add a library to your project, this does not mean that it will work. Be very careful because some calls will show themselves as non-functional only when called - no compile-time error will ever be thrown. Therefore it is a good idea to stress-test the library to the maximum before distributing it. When you are building libraries from source, make sure that all correct references are used (you can use the Class Library documentation to check this). Opinions expressed by DZone contributors are their own. {{ parent.title || parent.header.title}} {{ parent.tldr }} {{ parent.linkDescription }}{{ parent.urlSource.name }}
https://dzone.com/articles/not-every-library-will-work
CC-MAIN-2018-39
refinedweb
579
53.81
Plone 3.1.3 (Jul 16, 2008) This release is no longer supported! If you are using this release, please upgrade to a newer version if possible. Bugfix release for Plone 3.1.x For additional information about this project, please visit the project page . Available downloads For Windows (~ 26 Mb) For Linux (33MB) For all platforms (14 Mb) Release Notes The version of plone.app.customerize shipping with Plone 3.1.3 contains a bug preventing new view template customizations via /portal_view_customizations. Existing customizations will continue to work. To fix the problem an upgrade to version 1.1.2 is required. Buildout-based installations such as the one provided by the "Unified Installer" can easily be upgraded by making the following changes to buildout.cfg. For instances set up using the "Unified Installer" this file is located at zinstance/buildout.cfg relative to the target directory — usually /opt/Plone-3.1 or $HOME/Plone-3.1. The required changes are: A new [versions] section needs to be added, for example directly after the [buildout] section: [versions] plone.app.customerize = 1.1.2 In addition the line: eggs = plone.app.customerize needs to be added to the [plone] section. For a visual representation of these changes please refer to the ticket originally reporting the bug. After incorporating the changes buildout needs to be run again: $ bin/buildout -n The -n parameter should be used to ensure buildout is "Run in newest mode.". Change log - Updated translations. - Fix album display for folders which do not contain any images. This fixes 8212. - Fix invalid redirect in default view for comments. - Update the sharing page: - never reindex more than once when making changes. - modify the search button to only search. Previously it would also apply all changes in addition to doing a search. - do nothing if the form is submitted but no changes were made. - handle invalid LDAP or OpenID users correctly. - Update Archetypes: - use Content-Disposition instead of Content-disposition as HTTP header for downloads. Fixes problems with MS Internet Explorer 6, which appears to be case sensitive contrary to standards. - Add a new viewlet manager which allows adding of content at the bottom of content edit views. - Add an option to the reference field to also copy references when copying an object. - Update reference browser widget: - Make it possible to remove references for single-reference fields. - Add a new property 'startup_directory_method' to use a callable to determine the startup directory. - Inserted references were not checked by default on Internet Explorer 7. This fixes 7936. - Added a new property 'hide_inaccessible'. When set all items for which the user has no View permission are hidden. - Add a 'random items' option to the collection portlet. - Update GenericSetup import steps at each point when loading multiple profiles. Fixes CMF bug 213905. - Fix group handling to use group introspection PAS plugins, making it possible to ask, for example, LDAP groups for their members. - Fix user search to honour the order of user enumeration PAS plugins and correctly merge search results. - Add support for po files inside i18n folders in Python packages. They need to registered as a Zope2 product but don't need to be in the Products.* namespace anymore. - Do not create redirects for temporary URLs used while creating a new content object. This fixes 8260 and part of 7278. - Update viewlets to use the 'index' attribute rather than 'render' for setting viewlet templates. This allows them to be overriden using the 'template' ZCML attribute. - Do not show the display menu if it is disabled. - Add an actionMenuSelected class to selected menu items. - Standardize the rendering of the title in the folder contents template. - Fix a problem with disappearing security context for customized browser views. - Make it possible to customize viewlet and portlet templates registered with the same name for multiple browser layers. - Correctly handle objects with a space in their id in the broken-link-checker. Updated packages and products - Archetypes 1.5.9 - Products.ATReferenceBrowserWidget 2.0.2 - Products.CMFPlone 3.1.3 - Products.CMFQuickInstallerTool 2.1.6 - Products.GenericSetup 1.4.1 - Products.PlacelessTranslationService 1.4.12 - Products.PlonePAS 3.6 - Products.PloneTranslations: 3.1.3 - five.customerize 0.3 - plone.app.content 1.1.3 - plone.app.contentmenu 1.1.3 - plone.app.customerize 1.1.1 - plone.app.layout 1.1.3 - plone.app.linkintegrity 1.0.10 - plone.app.portlets 1.1.3 - plone.app.redirector 1.0.9 - plone.app.workflow 1.1.3 - plone.fieldsets 1.0.2 - plone.portlet.collection 1.1.3 - plone.session 2.0
http://plone.org/products/plone/releases/3.1.3
crawl-002
refinedweb
761
51.04
On 01/19, Cyrill Gorcunov wrote:>> On Thu, Jan 19, 2012 at 07:51:12PM +0400, Cyrill Gorcunov wrote:> > If it's needed I can wrap all this with CONFIG_CHECKPOINT_RESTORE, should I?> >> > --->> Oleg, if only I'm not missing something obvious you meant handling like below?Yes, but...> +struct proc_pid_children_iter {> + struct pid_namespace *pid_ns;> + struct pid *parent_pid;> +};you forgot to remove this definition.> +static int children_seq_show(struct seq_file *seq, void *v)> +{> + struct inode *inode = seq->private;> + unsigned long pid;> +> + pid = (unsigned long)pid_nr_ns(v, inode->i_sb->s_fs_info);> + return seq_printf(seq, " %lu", pid);> +}just noticed... why unsigned long and %lu? afaics pid_t/%d should workwithout any typecasts.Oleg.
http://lkml.org/lkml/2012/1/19/309
CC-MAIN-2017-30
refinedweb
106
56.05
I'm sending SNS with Ruby Amazon SDK. As far I understood, the method Aws::SNS::Client#publish When a messageId is returned, the message has been saved and Amazon SNS will attempt to deliver it to the topic's subscribers shortly. publish You can query for topics or for subscriptions, but you can't query for published messages. I doubt AWS keeps them, but if they do, there is no API to access them so far as I know. One workaround would be to subscribe to each of your own topics using an http or https endpoint on your own site. Each time you create a topic, you would create a corresponding subscription to your own endpoint: def generate_sns_topic(topic_name) sns_client = Aws::SNS::Client.new response = sns_client.create_topic(name: topic_name) if response.successful? sns_client.subscribe(topic_arn: topic_response.topic_arn, protocol: :https, endpoint: <your_site_endpoint>) else <error_handling_here> end end Now you need to create POST <your_site_endpoint> to accept the AWS message as it comes in. Once you receive it, you'll know SNS has sent it. Presumably you'll create a new published_messages table in your database to keep track of what's been published.
https://codedump.io/share/kk4OoQ7YMZ9I/1/amazon-sns-sending-confirmation-with-ruby-sdk
CC-MAIN-2017-47
refinedweb
193
63.9
Provided by: plplot-doc_5.10.0+dfsg2-0.1ubuntu2_all NAME plsmin - Set length of minor ticks SYNOPSIS plsmin(def, scale) DESCRIPTION This sets up the length of the minor ticks and the length of the terminals on error bars. The actual length is the product of the default length and a scaling factor as for character height. Redacted form: plsmin(def, scale) This function is used in example 29. ARGUMENTS def (PLFLT, input) The default length of a minor tick in millimeters, should be set to zero if the default length is to remain unchanged. scale (PLFLT, input) Scale factor to be applied to default to get actual tick length. AUTHORS Many developers (who are credited at) have contributed to PLplot over its long history. SEE ALSO PLplot documentation at. February, 2016 PLSMIN(3plplot)
http://manpages.ubuntu.com/manpages/xenial/man3/plsmin.3plplot.html
CC-MAIN-2019-43
refinedweb
134
65.83
Summary: One of the following methods can be used to check if a list is empty :- - Method 1: Using isinstance() With any() - Method 2: Using isinstance() And len() Methods Within For Loop Problem: Given a list; how to check if it is nested or not? Considering that you have a list such that it is nested at times while sometimes it is not nested. Based on whether it is nested or not the continuation of your program flow would be different. For instance, you are feeding the list into a dataframe. The code for doing that is different depending on whether the list is flat or nested. Flattening the nested list yields a different structure to the data. Thus, you need to maintain the structure. So, how would you differentiate and recognize a nested list from a flattened list (non-nested list)? Example: [a,b,c] # Output: --> False [[1,2,3],[4,5,6]] # Output --> True In this article let us quickly discuss the methods that can be used to check if a given list is nested or not. So, without further delay let us dive into the solutions. Method 1: Using isinstance() With any() The easiest solution to our problem is to use the isinstance() method and a generator expression within the any() function. Before diving into the solution, let us understand the usage of the isinstance() and any() methods that will help us to check the list if it is nested or not. ◉ isinstance is a built-in method in Python which returns True when the specified object is an instance of the specified type, otherwise it returns False. Syntax: Example: a = isinstance(25, int) print(a) Output: True ◉ any() is a built-in function that returns True if any element in an iterable is True, otherwise it returns False. Syntax: any(iterable) Example: li = [0, 10, 100, 1000] x = any(li) # Returns True because the first item is True print(x) Output: True Now that we know the usage of each function, let us have a look at the solution to our problem. Please follow the code given below that demonstrates the solution. Solution li_1 = [1, 2, 3] # flat list li_2 = [[1, 2, 3], [4, 5, 6]] # nested list # Testing for nested list li_1_output = any(isinstance(i, list) for i in li_1) li_2_output = any(isinstance(i, list) for i in li_2) # Printing output print("Is li_1 Nested?", li_1_output) print("IS li_2 Nested?", li_2_output) Output: Is li_1 Nested? False IS li_2 Nested? True Explanation In the above code the any() method allows us to check for each occurrence of the list while the isinstance() method checks if each instance of an element inside the list is a list itself or not. Therefore in the first case, the output is False since Python does not find any occurrence of another list within li_1 while in the second case it finds a couple of lists within the parent list li_2 and it returns True. Method 2: Using isinstance() And len() Methods Within For Loop Another work around for our problem is to use a counter variable that counts the number of elements within the list and compare it with the actual length of the list. If the length of the list is equal to the number of elements within the list then it is not a nested list otherwise it is a nested list. Let us have a look at the program given below that demonstrates this concept. (Please follow the comments along with the code for a better understanding.) li_1 = [1, 2, 3] # flat list li_2 = [[1, 2, 3], [4, 5, 6]] # nested list # function to evaluate if the list is nested or not def count(l): flag = 0 # counter variable to keep count of the number of elements in the list # iterate through the elements of the list for item in l: # check if the item is a list (iterable) or not if isinstance(item, list): flag = flag + len(item) else: flag = flag + 1 return flag x = count(li_1) print("Is li_1 Nested? ", bool(x != len(li_1))) y = count(li_2) print("Is li_2 Nested? ", bool(len(li_2) != y)) Output: Is li_1 Nested? False Is li_2 Nested? True The NumPythonic Way Another interesting approach could be to convert the given list into an array using the Numpy library. Further, if you are working with datasets then it is quite possible that you are dealing with arrays instead of lists which makes it even more reasonable to discuss the NumPythonic way of approaching our problem. While dealing with arrays our task becomes extremely easy because all we have to do now is to check if the given array is a one-dimensional array. #Note: The ndim attribute allows us to find the dimension of an array and the size() method allows us to find the number of elements of an array along a given axis. Let us have a look at the following program which will help us to understand how easily we can deduce if the given list is nested or not by converting it to an array. import numpy as np li_1 = [[1, 2, 3, 4], [5, 6, 7, 8]] li_2 = [1, 2, 3, 4, 5] arr_1 = np.array(li_1) arr_2 = np.array(li_2) print("Is arr_1 nested? ", bool(arr_1.ndim > 1)) print("Is arr_2 nested? ", bool(arr_2.ndim > 1)) Output: Is arr_1 nested? True Is arr_2 nested? False Edit: If you wish to learn NumPy based on puzzle-based learning then you might be fascinated to have a look at this amazing book published by Chris who helps you to learn with the help of puzzles in this book. Do you want to become a NumPy master? Check out our interactive puzzle book Coffee Break NumPy and boost your data science skills! (Amazon link opens in new tab.) Conclusion I hope the methods discussed in this article have helped you to learn some very basic yet very important concepts that can be used to verify if a list is nested or not. The major concepts discussed in this article were: - Using isinstance()and any()methods to verify if the given list is nested? We also discussed each method in brief before reaching the final solution. - Using isinstance()And len()Methods Within For Loop to verify if the given list is nested? - Using the numpylibrary to verify if the given list is nested? With that we come to the end of this article;!
https://blog.finxter.com/how-to-check-if-a-list-is-nested-in-python/
CC-MAIN-2021-43
refinedweb
1,075
67.99
socket connection - $30 rewardEladElrom Feb 19, 2007 6:39 AM I am building an application that uses the socket connection. When the application is running from my desktop it is working fine, however when it's running on a remote server I am getting sandbox security violation: "cannot load data from serverName.com:portNumber.". This content has been marked as final. Show 22 replies 1. Re: socket connection - $30 rewardmichael_ramirez44 Feb 19, 2007 7:17 AM (in response to EladElrom)What are you trying to connect to? The server you are trying to connect to must have a crossdomain.xml file defined that allows you to have access unless it is from the same domain that the swf was loaded from. 2. Re: socket connection - $30 rewardEladElrom Feb 19, 2007 8:08 AM (in response to michael_ramirez44)It has the crossdomain.xml which is set to "*", which should allow any server connection. The application doesn't work without the crossdomain.xml. 3. Re: socket connection - $30 rewardFlightGuy Feb 19, 2007 9:09 AM (in response to EladElrom)Are you loading the .swf from? or is it being loaded from a different server (and do you need to load it from a different server?) Tim 4. Re: socket connection - $30 rewardFlightGuy Feb 19, 2007 9:13 AM (in response to EladElrom)You won't be able to make the AIM application work from anywhere other than the desktop because the crossdomain.xml needs to be in the AIM server (and it's not likely you'll persuade them - unless you work there - to post a crossdomain.xml file). Something I'm not sure of, and I'd be interested in hearing from someone who is, is how Flash determines the url to search for the crossdomain.xml file. Does it assume it's port 80? and would it, in and example such as this, look at? If so, and if you haven't done so yet, deploy your crossdomain.xml file to the root of. Tim 5. Re: socket connection - $30 rewardEladElrom Feb 19, 2007 9:16 AM (in response to EladElrom)I tried it both ways. Putting it in the same site URL and different server both didn't worked. the swf is only working on the desktop. At the end the swf file need to connect to a different server, but I just want to see it working, for now.... 6. Re: socket connection - $30 rewardEladElrom Feb 19, 2007 9:25 AM (in response to EladElrom)Thanks Tim. I am trying to get my application to work and not the AOL AIM. I used the AIM as an example so people can understand what I am talking about but I know it lack the XML file and will not work without it... I did put the XML on the server with permission to let access to all servers like this: <allow-access-from so it should have worked. However it only works when the application is running on the desktop. 7. Re: socket connection - $30 rewardFlightGuy Feb 19, 2007 9:43 AM (in response to EladElrom)OK, I read the doc - you'll find the relevant sections on page 468 of the "Programming ActionScript 3.0" manual. You have some options about how to deliver the socket cross-domain policy file. You can deliver it from the same port (5190) by making the socket server respond to the request <policy-file-request/> by sending the contents of your crossdomain.xml file (with the to-ports="5190" property set). After doing this, end the connection - flex will establish a new connection once it's loaded the policy file. Alternatively you can listen on another socket for policy requests, in which case a connection to that port can be answered immediately by sending the contents of the crossdomain.xml file. In this case, before you connect to port 5190, call: Security.loadPolicyFile("xmlsocket://aimexpress.oscar.aol.com:5151"); // or whatever port you've chosen As far as I can tell you can also use a normal HTTP delivered crossdomain.xml file, and if you're using a port above 1024 you don't need to specify the to-port property in the crossdomain.xml file. In this case it would need to be accessible as. In this case, you will need to explicitly load the policy file as well: Security.loadPolicyFile(""); I haven't tried any of this, but will need to in due course, so I would appreciate feedback if this works. Tim 8. Re: socket connection - $30 rewardFlightGuy Feb 19, 2007 9:44 AM (in response to EladElrom)What port number are you trying to connect to? If it's below 1024 I don't believe you can use and http served policy file, and you'll need to use one of the two socket options. Tim 9. Re: socket connection - $30 rewardEladElrom Feb 19, 2007 9:49 AM (in response to EladElrom)I heard about that before... the only problem is that my application is using port 80, which is the lower port and I am not sure if it will work, but I will try to make the AOL application to work and that can be a good start... I will let you know... 10. Re: socket connection - $30 rewardEladElrom Feb 19, 2007 9:51 AM (in response to EladElrom)Also.. .I am actually using the "Security.loadPolicyFile" in my application and it didn't helped in any way... 11. Re: socket connection - $30 rewardEladElrom Feb 19, 2007 10:02 AM (in response to EladElrom)I just tried using the: import flash.system.Security; Security.loadPolicyFile("xmlsocket://aimexpress.oscar.aol.com:5151"); on the AIM application, no luck there... 12. Re: socket connection - $30 rewardFlightGuy Feb 19, 2007 11:03 AM (in response to EladElrom)My choice of 5151 in the AOL example was arbitrary - you would need to have code running on that machine to respond to the connection. You say you're using port 80 in your code, and using a socket connection - XMLSocket, or Socket? If you want to make a socket connection to port 80 you can't use an HTTP-delivered crossdomain.xml file. Do you have enough freedom in your protocol to support the same-socket crossdomain service? ie. can you make your server send no response to the client until it receives data from the client, and then if the received data is <policy-file-request/> then answer with the contents of your crossdomain file? As I mentioned before, unless you're free to add code to AOL's server, you cannot make the AIM example work. The whole purpose of the crossdomain file is to require the provider of the service (AOL in this case) to authorize the access, and the only way to do this is to get them to post a crossdomain file - either via http (allowing access to all ports above 1024 and nothing below), or through a socket service (allowing access to all nominated ports). Can you explain a little more about what you're trying to do, what context you expect it to be used, and what your reason is for using a direct socket connection on port 80 - nothing wrong with it, just unconventional. I'm guessing you're using port 80 to avoid firewall issues (which, if you're not using http protocol does only 90% the job). This would help me understand whether you'd do better to use the same-socket option or have a dedicated socketserver to handle the policy file. Tim 13. Re: socket connection - $30 rewardEladElrom Feb 19, 2007 12:17 PM (in response to EladElrom)My class waits for a handshake before requesting information, I am forced to use low port (you were right; I need to use low port for security reasons). After the handshake the two servers are exchanging information using an event manager. I prefer to use the binary connection since there is a lot of information to pass between the servers, and I know that I can do it in different ways such as soap etc.. But I really would like to get this to work. The error occurs from the handshake. I was trying both; let the server connect to itself, or connect a different server, unfortunately both cases didn't work. I am really not that concern to get AIM example to work, since I was just using it as an example. I prefer not to post my script on an open form like this. I realize that it was the wrong example since the AIM script need the XML which will not going to happened without AOL posting the xml file on their server. Tim, are you interested helping me out with this error and I am willing to pay you for consulting on hourly rate or on a fixed rate? 14. Re: socket connection - $30 rewardFlightGuy Feb 19, 2007 1:50 PM (in response to EladElrom)Hi - I'm not interested in consulting for this or accepting payment, but I'm happy to work with you on the forum to get it sorted out. 15. Re: socket connection - $30 rewardFlightGuy Feb 19, 2007 2:32 PM (in response to EladElrom)Given that you're using port 80 and are limited to use this port to get past firewalls, you really have only one option - you cannot use http to serve the policy file, since you're already using port 80 for your socket connection. You cannot use a different port to serve it, since that would defeat your firewall solution. This leaves you with only the option of serving the policy file through the same port. You say that your server script waits for a handshake before starting to communicate. Can you make your server script, upon connection, differentiate between your client-to-server handshake message and the <policy-file-request/>? If so, when you detect the <policy-file-request/>, simply reply with the following text, and then close the socket: <cross-domain-policy><allow-access-from</cross-domain-policy> If you do this, you shouldn't even need to do anything in your flex code. The first time a client connection is requested, flash will first connect to this socket and send <policy-file-request/>. It will wait for a reply in the form of a cross-domain-policy document, and then disconnect the socket connect. Having received this (if it allows connection), flash will reconnect with your Socket or XMLSocket connection and you can then use your own protocol. If there's some script or code that you have that you feel would help explain the issues you're having, you can send me a private message and I'll have a look. Tim 16. Re: socket connection - $30 rewardEladElrom Feb 19, 2007 2:34 PM (in response to EladElrom)Ok, thanks. Thank god for this forum and adobe help... The closest to what I am doing is FTP connection, so I will create a very simple script and post it soon. 17. Re: socket connection - $30 rewardFlightGuy Feb 19, 2007 2:40 PM (in response to EladElrom)By the way using http still gives you access to a binary stream in each direction once you've established the connection. You can transfer substantial volumes of compressed binary data. I used to do this, but then the need arose to be more responsive on the client when events happen on the server. This led me away from an http polling model to a socket connection. The socket connect is great because it allows a server-side push of data. If you are going to need to send unsolicited data from the server to the client, it's a good option. However if you're just going to be transferring large volumes of data in response to a client request, you would do well to stick with http with next to no protocol overhead (unless you're sending several scores of requests per minute, in which case it becomes a bit chatty). Tim 18. Re: socket connection - $30 rewardEladElrom Feb 20, 2007 6:32 AM (in response to FlightGuy)Hi Tim, I am embarrassed to admit :-( , however the flash movie was not pointing to the right crossdomain.xml file…It is working fine now... Thanks so much for your help, I REALLY REALLY appreciate all your responses and help yesterday. PS: Are you working for Adobe? 19. Re: socket connection - $30 rewardFlightGuy Feb 20, 2007 2:35 PM (in response to EladElrom)You're welcome - no I don't work for Adobe. I'm just using flex for a project and it's in my interest for Flex to be a successful product and to have a reputation as reliable and solid. I'm doing what I can, Tim 20. Re: socket connection - $30 rewardEladElrom Feb 22, 2007 3:34 PM (in response to EladElrom)The cross XML file was placed in the wrong place, it is now working correctly. 21. Re: socket connection - $30 rewardKarpagarajan May 21, 2007 6:54 AM (in response to EladElrom)Hi... I am facing the same problem. Can you tell me the correct path for that crossdomain.xml? thanks 22. Re: socket connection - $30 rewardmirkasim Jun 1, 2008 11:11 AM (in response to EladElrom)Hi, go to Flash Resources , you can find a java application that can serve policy files to resolve this problem.
https://forums.adobe.com/thread/172473
CC-MAIN-2019-04
refinedweb
2,244
59.43
0 Hello, When random generates an output, it cannot show duplicate strings. So for example the output I am getting (randomly) is, "Freddy, Freddy, Jane" when it should be "Freddy, Jane". I have to hardore so I cannot use Sets or ArrayList, Contains or for anything like that. Also no Stringbuilder. I don't know how to fix this. Does anybody know to kindly help me. Thanks in advance java.util.Random. public class frogs { private static Random nums = new Random(); private static String[] names = { "Freddy", "Mac", "Jane", "Jule" }; public static int Dice() { return (names.nums.nextInt(2) + 1); } public String randomNames() { String tempString = " "; int numOfTimes = Dice(); int dup = 0; for(int i=0 ; i<numOfTimes; i++) { if (names[i].equals(names[i])) [I]// if two strings are the same then it should //not output the string again. // I know this loop is wrong but I don't how to //fix this [/I] { i = 0; } tempString = tempString + names[nums.nextInt(names.length)]; tempString = tempString + ","; } return aTemp; } } } Edited by peter_budo: Keep It Organized - For easy readability, always wrap programming code within posts in [code] (code blocks)
https://www.daniweb.com/programming/software-development/threads/290544/help-on-duplicate-arrays-the-hardcore-way
CC-MAIN-2017-26
refinedweb
185
65.83
Over the past few years, there has been an explosion in the number of web sites dedicated to teaching programming. A few, like Codecademy, are making an effort to offer multilingual content. While there seems to exist a prejudice shared by many programmers that "real programmers code in English and therefore everyone should learn to program using English", there is no doubt that for most non-English speakers, having to learn programming concepts and new English vocabulary at the same time can make the learning experience more challenging. So, for beginners, the best learning environment is one that is available in their native tongue. However, to do so can require a massive amount of work and a team of people. Freely available programming environments rely on the help of volunteers to provide translation. Viewed from the outside, the work required to translate a programming environment like that of codecademy appears to be an all or nothing kind of effort, as everything (UI, lessons and feedback messages) is tightly integrated. A given tutorial is broken up in a series of lessons, each of which can be translated independently. For many of these lessons, the range of acceptable input by users is fairly limited, and so are the possible feedback messages. Still, the amount of work required for providing a complete translation of a given tutorial is enormous. When I started the work required to create Reeborg's World, I knew that I would be largely on my own. Still, I wanted to make sure that the final "product" would be available in both English and French, and could be adapted relatively easily for other languages. The approach I have used for Reeborg's World is different than that of codecademy, and has been greatly influenced by my past experiences with RUR-PLE and Crunchy, as well as during the development of Reeborg's World itself. 1. I have completed separated the tutorials from the programming environment itself. Thus, one can find a (slightly outdated) tutorial for complete beginners in English as well as a French version of the same tutorial, separately from the programming environment itself. This separation of the tutorial and the programming environment makes it easily possible for others to create their own tutorial (in whatever language) like this beautiful tutorial site created by a teacher in California that makes use of Reeborg's World, but calls the robot Karel as a hat tip to the original one created by Richard Pattis. 2. The programming environment is a single web page, with a mimimum amount of visible text. Currently, there are two versions: one in English, and one in French. 3. Programming languages available are standard ones (currently Python, Javascript and CoffeeScript). Whereas I have seen translated "real" programming language (like ChinesePython or Perunis, both of which are Python with keywords and builtins translated in other languages) or translated mini-programming language, like that used by Guido van Robot, I do not believe that the slight additional work required to memorize the meaning of keywords and builtins is a significant hurdle, especially when taking into account the possibilities of using the programming language outside of the somewhat artificial learning environment. 4. Possible feedback given to the user are currently available in both French and English. The approach I have used is to create a simple Javascript function RUR.translate = function (s) { if (RUR.translation[s] !== undefined) { return RUR.translation[s]; } else { return s; } }; Here, RUR is a namespace that contains the vast majority of functions and other objects hidden from the end user. A given html page (English or French) will load the corresponding javascript file containing some translations, such as RUR.translation["Python Code"] = "Code Python"; The total number of such strings that need to be translated is currently slightly less than 100. 5. Functions (and a class) specific to Reeborg's World are available in both English and French, as well as in Spanish (for the Python version). By default, the English functions are loaded on the English page and the French ones are loaded on the French page. An example of a definition of a Python command in English could be: move = RUR._move_ class UsedRobot(object): def move(self): RUR.control.move(self.body) whereas the French equivalent would be avance = RUR._move_ class RobotUsage(object): def avance(self): RUR.control.move(self.body) with similar definitions done for Javascript. 6. When using Python, one can use commands in any human language by using a simple import statement. For example, one can use the French version of the commands on the English page as follows: from reeborg_fr import * avance() # equivalent to move() move() # still works by default If anyone is interested in contributing, this would likely be the most important part (and relatively easy, as shown above) to translate in other languages. 7. As part of the programming environment, a help button can be clicked so that a window shows the available commands (e.g. move, turn_left, etc.). When importing a set of commands using Python in a given human language as above, there is also a provision to automatically update the help available on that page based on the content of the command file. (While the basic commands are available in Spanish, the corresponding help content has not been translated yet.) 8. In order to teach the concept of using a library, two editor tabs are available on the page of the programming environment, one of which represents the library. In the tutorial I wrote, I encourage beginners to put the functions that they define and reuse in many programs, such as turn_right or turn_around, in their library so that they don't have to redefine them every single time. The idea of having a user library is one that was requested in the past by teachers using RUR-PLE. The idea of using a second editor on the second page is one that I first saw on this html canvas tutorial by Bill Mill. When programming in Javascript (or CoffeeScript), for which the concept of a library is not native like it is in Python, if the user calls the function import_lib(); the content of the library will be evaluated at that point. For Python, I ensure that the traditional way works. Brython (which I use as the Python in the browser implementation) can import Python files found on the server. For the English version, the library is called my_lib.py and contains the following single line: from reeborg_en import * (The french version is called biblio.py and contains a similar statement.) When a user runs their Python program, the following code is executed def translate_python(src): import my_lib #setup code to save the current state of my_lib exec(library.getValue(), my_lib.__dict__) exec("from reeborg_en import *\n" + src) # cleanup to start from a clean slate next time import my_lib #setup code to save the current state of my_lib exec(library.getValue(), my_lib.__dict__) exec("from reeborg_en import *\n" + src) # cleanup to start from a clean slate next time The content of the editor (user program) is passed as the string src. library.getValue()returns the content of the user library as a string. Even if my_lib only imported once by Brython, its content effectively gets updated each time a program is run by a user. 9. One thing which I did not translate, but that we had done for Crunchy (!), are the Python traceback statements. However, they are typically simplified from the usual full tracebacks provided. In summary, to provide a complete translation of Reeborg's World itself in a different language, the following are needed: 1. a translated single page html file; 2. a javascript file containing approximately 100 strings used for messages; 3. a Python file containing the defintion of the robot functions as well as a corresponding "help" section; 4. a javascript file containing the definition of the robot functions. If the Python file containing translation is available, it is trivial for me to create this corresponding javascript file.
https://aroberge.blogspot.com/2014/11/translating-programming-environment.html
CC-MAIN-2018-13
refinedweb
1,340
59.43
Log4net is an open source library that allows .NET applications to log output to a variety of sources (e.g., The console, SMTP or files). Log4net is a port of the popular log4J library used in Java. Full details of log4net can be found at its project homepage. In this article, I will give an overview of how the library works and examples of a few of the different logging options. The log4net library can be downloaded from the project homepage. The latest distribution comes with versions of log4net for MONO, .NET 1.0 and 1.1, and .NET Compact Framework. Log4net provides a simple mechanism for logging information to a variety of sources. Information is logged via one or more loggers. These loggers provide 5 levels of logging: Hopefully (!!) the amount of logging performed by each level will decrease going down this list (we want more debug information than fatal errors!). The level of logging on a particular logger can be specified, hence during development, all 5 levels of logging can be output to aid development. When your application is deployed, you may decide to only output level 5 – Fatal errors to your log files. It's up to you and it's all easily configurable. So, where does the logging information go? Well, logging information goes to what is called an Appender. An appender is basically a destination that the logging information will go to. In this article, I will give examples of the ConsoleAppender and the FileAppender. Many other appenders exist to allow data to be logged to databases, email, net broadcasts etc. You are not limited to only using one appender, you can have as many appenders configured for use as you want. Appender configuration is performed outside of code in XML files, so as we shall see later, it is a simple matter to change your logging configuration. OK, so we know that we can use a logger to output data to a number of appenders, but what format will the information be logged as? Well, log4net has a number of layouts that can be used for each appender. These layouts specify whether the logs are produced as simple textual data or as XML files, or whether they have timestamps on them etc. OK, I'm sure you've read enough about logging now, and want to see some code and see how it all works. Well, here goes. Our first (incredibly simple) example program (LogTest.exe) is shown below: using log4net; using log4net.Config; public class LogTest { private static readonly ILog logger = LogManager.GetLogger(typeof(LogTest)); static void Main(string[] args) { BasicConfigurator.Configure(); logger.Debug("Here is a debug log."); logger.Info("... and an Info log."); logger.Warn("... and a warning."); logger.Error("... and an error."); logger.Fatal("... and a fatal error."); } } You can see that there are a few interesting lines within this small application. The first of these is where we create a Logger. private static readonly ILog logger = LogManager.GetLogger(typeof(LogTest)); This creates a logger for the class LogTest. You don't have to use a different logger for each class you have, you can use different loggers for different sections or packages within your code. The class of the logger is output to the log so you know where any logged information has come from. LogTest The next interesting line is: BasicConfigurator.Configure(); This method initializes the log4net system to use a simple Console appender. Using this allows us to quickly see how log4net works without having to set up different appenders. Don't worry however, setting up appenders isn't difficult and we'll look at that in a minute. Finally, if you run the application, you can see the different levels of logging output that are performed - in this case, to the console: 0 [2436] DEBUG LogTest - Here is a debug log. 31 [2436] INFO LogTest - ... and an Info log. 31 [2436] WARN LogTest - ... and a warning. 31 [2436] ERROR LogTest - ... and an error. 31 [2436] FATAL LogTest - ... and a fatal error. Let's now change the test application so that it loads the log4net configuration information from a “.config” file. This is done by removing the BasicConfigurator and using a DOMConfigurator as is shown in the sample application LogTest2. BasicConfigurator DOMConfigurator using log4net; using log4net.Config; public class LogTest2 { private static readonly ILog logger = LogManager.GetLogger(typeof(LogTest2)); static LogTest2() { DOMConfigurator.Configure(); } static void Main(string[] args) { logger.Debug("Here is a debug log."); logger.Info("... and an Info log."); logger.Warn("... and a warning."); logger.Error("... and an error."); logger.Fatal("... and a fatal error."); } } Now that we are using a DOMConfigurator, we need to specify our appenders and layouts. This is done in the application's .config file, “LogTest2.exe.config”. <?xml version="1.0" encoding="utf-8" ?> <configuration> <configSections> <section name="log4net" type="log4net.Config.Log4NetConfigurationSectionHandler, log4net" /> </configSections> <log4net> <appender name="LogFileAppender" type="log4net.Appender.FileAppender"> <param name="File" value="LogTest2.txt" /> <param name="AppendToFile" value="true" /> <layout type="log4net.Layout.PatternLayout"> <param name="Header" value="[Header]\r\n" /> <param name="Footer" value="[Footer]\r\n" /> <param name="ConversionPattern" value="%d [%t] %-5p %c %m%n" /> </layout> </appender> <appender name="ConsoleAppender" type="log4net.Appender.ConsoleAppender" > <layout type="log4net.Layout.PatternLayout"> <param name="Header" value="[Header]\r\n" /> <param name="Footer" value="[Footer]\r\n" /> <param name="ConversionPattern" value="%d [%t] %-5p %c %m%n" /> </layout> </appender> <root> <level value="INFO" /> <appender-ref <appender-ref </root> </log4net> </configuration> Within this XML file, you can see that there are two appenders specified, LogFileAppender and ConsoleAppender. We also have a root appender. This specifies what appenders we are using and also what level of output we want to see in our logs – in this case, everything from INFO onwards (i.e., everything except DEBUG). The LogFileAppender specifies which file to use and whether to append onto the end of an existing file or not. Both appenders use a common layout that specifies what information is output to the logs. We can see that the date (%d), time (%t), logging level (%p), the name of the logger (%c) and the message (%m) are output. If you now run application LogTest2, the following output is shown on the console: 2004-09-10 12:53:48,062 [1216] INFO LogTest2 - ... and an Info log. 2004-09-10 12:53:48,078 [1216] WARN LogTest2 - ... and a warning. 2004-09-10 12:53:48,078 [1216] ERROR LogTest2 - ... and an error. 2004-09-10 12:53:48,078 [1216] FATAL LogTest2 - ... and a fatal error. If you have a look at the file “LogTest2.txt”, you will see that it contains the same information. Well, that's a brief overview of what log4net can do, and we've only really scratched the surface here. Log4net provides many other features and I'd recommend grabbing hold of a copy to try them.
http://www.codeproject.com/Articles/8245/A-Brief-Introduction-to-the-log4net-logging-librar?fid=103589&df=90&mpp=10&sort=Position&spc=None&select=4464838&tid=4157597
CC-MAIN-2017-17
refinedweb
1,151
60.61
Newlines in Sources.bz2 indices Bug Description Hi This looks similar to bug #435316, but I'm not sure it's the same. In http:// [...] Package: openoffice.org Binary: openoffice.org, broffice.org, openoffice. libuno- Version: 1:3.1.1- [...] there's a superfluous newline after libuno- [ This is breaking apt_pkg's parsing of this file which breaks germinate for the Ubuntu Moblin Remix image so it's quite high on my radar right now. ;-) ] Related branches - Colin Watson: Approve on 2009-10-15 - Gavin Panella (community): Approve on 2009-10-07 - Diff: 537 lines9 files modifiedlib/canonical/launchpad/emailtemplates/upload-accepted.txt (+1/-0) lib/canonical/launchpad/emailtemplates/upload-announcement.txt (+1/-0) lib/canonical/launchpad/emailtemplates/upload-new.txt (+1/-0) lib/canonical/launchpad/emailtemplates/upload-rejection.txt (+1/-0) lib/lp/archiveuploader/changesfile.py (+15/-1) lib/lp/archiveuploader/tagfiles.py (+57/-22) lib/lp/archiveuploader/tests/data/test436182_0.1_source.changes (+23/-0) lib/lp/archiveuploader/tests/test_tagfiles.py (+101/-37) lib/lp/soyuz/model/queue.py (+15/-7) It seems we have accepted some DSC with this format mistake: {{{ launchpad_prod=> SELECT spn.name as name, spr.version as version, o.name as owner, a.name as archive from sourcepackagere %'; name | version | owner | archive ------- openoffice.org | 1:3.1.1-2ubuntu1 | ubuntu-drivers | primary linux | 2.6.31- mono | 2.4.2.3+dfsg-2 | ubuntu-drivers | primary linux | 2.6.31- openoffice.org | 1:3.1.1- linux | 2.6.31-10.35~awe1 | awe | ppa linux | 2.6.31-10.36~eee1 | yofel | ppa linux | 2.6.31- linux | 2.6.31-10.36~eee2 | yofel | ppa linux | 2.6.31-10.36~eee3 | yofel | ppa linux | 2.6.31- openoffice.org | 1:3.1.1- (12 rows) This is somehow related with bug #435316, we have to find the right way to identify this format problem (I'm assuming it is a problem) and reject those uploads. Since there are not *many* yet, the migration of the existing that will be easy. From the Debian (and Ubuntu) Policy Manual: 5.6.19. `Binary' ---------------- This field is a list of binary packages. Its syntax and meaning varies depending on the control file in which it appears. When it appears in the `.dsc' file, it lists binary packages which a source package can produce, separated by commas[1]. It may span multiple lines. This is preventing OEM Services from running germinate for some Intel work we're doing. Very important this gets fixes ASAP. Just trying to reproduce this locally - with a pristine sources.txt in the same directory in a python console: {{{ f = open('sources.txt') import apt_pkg p = apt_pkg. while p.Step() == 1: ver = p.Section[ src = p.Section[ f.close() }}} Add *two* consecutive new-lines within the Binary: section as you like and save, then: {{{ f = open('sources.txt') p = apt_pkg. while p.Step() == 1: src = p.Section[ ver = p.Section[ }}} The two consecutive new-lines results in: Traceback (most recent call last): File "<console>", line 3, in ? KeyError: 'Version' But using just one new-line (as described in the bug) seems to work fine. This is with apt_pkg.Version == '0.7.20.2ubuntu6'. I'll try to find out which version of apt-pkg is being used and verify the problem. OK, exactly the same (two new-lines required) to replicate this on dogfood which has: $ dpkg-query -W apt python-apt apt 0.7.9ubuntu17. python-apt 0.7.4ubuntu7.5 which, after checking with cjwatson is the same as the production publisher. So it seems when the original dsc contains a single \n, we're publishing a Sources.bz2 with a *blank* line, which is incorrect and apt_pkg correctly fails to parse it. OK, now for the fix... The actual issue is in lp.archiveuploa I've created and tested an intermediate work-around that rstrips the dsc_binaries when creating the index file, which will give us a bit more time to solve the original issue properly. I'm setting the status back to Triaged, as what we've committed is a work-around - not a fix. It's a work-around that we can cherry-pick if necessary, and it will ensure that there are no trailing '\n' at end of fields in the Sources index file, so that the file can be correctly parsed by apt_pkg. The work-around does not ensure that the stray trailing '\n' will not get into the SPR.dsc_binaries field in the first place, nor does it fix the corrupt data. Hence leaving this bug open. I've not verified it yet, but we're 90% certain that when there are valid single '\n' chars within a field's values (as in the snippet in the bug description above), the lp.archiveuploa I can't see any reason why we are not simply using apt_pkg. As another example, currently the Sources in the following ppa (identified by Celso's query above) displays the problem also: http:// {{{ Package: linux Binary: linux-source- linux-image- Downloading http:// ppa.launchpad. net/moblin/ ppa/ubuntu/ dists/karmic/ main/source/ Sources. bz2 file ... ppa.launchpad. net/moblin/ ppa/ubuntu/ dists/karmic/ main/source/ Sources. bz2 file ... germinate- update- metapackage" , line 273, in <module> germinate/ Germinate/ Archive/ tagfile. py", line 121, in feed Sources" )) germinate/ Germinate/ germinator. py", line 323, in parseSources "Version" ] Decompressing http:// Traceback (most recent call last): File "/usr/bin/ germinator, [dist], components, architecture, cleanup=True) File "/usr/lib/ "source/ File "/usr/lib/ ver = p.Section[ KeyError: 'Version'
https://bugs.launchpad.net/launchpad/+bug/436182
CC-MAIN-2021-21
refinedweb
922
61.22
$ cnpm install n3tr-next Next.js is a minimalistic framework for server-rendered React applications. Visit to get started with Next.js. <head> <Link> <Link> <Document> Install it: The beta has support for the latest version of React (v16) and is actively being developed upon. npm install next@beta react react-dom This is the stable version of Next.js npm install next react@15 react-dom@15 --save: . <Link>: getInitialProps, data is fetched. If an error occurs, _error.jsis rendered pushStateis performed and the new component" /> </Link> </div> hrefto. import Link from 'next/link' import Unexpected_A from 'third-library' export default ({ href, name }) => <Link href={href} passHref> <Unexpected_A> {name} </Unexpected_A> </Link> You can also do client-side page transitions using the next/router import Router from 'next/router' export default () => <div> Click <span onClick={() => Router.push('/about')}>here</span> to read more </div>) download JS code. When the page is getting rendered, you may need to wait for the data. > Typically you start your next server with next start. It's possible, however, to start a server 100% programmatically in order to customize routes, use route patterns, etc This example makes /a resolve to ./pages/b, and /b resolve to ./pages(path: string, opts: object)- pathis where the Next project is located.: // .. 404 or 500 errors are handled both client and server side by a default component error.js. If you wish to override it, define a _error.js in the pages folder:> ) } } If you want to render the built-in error page you can by using next/error: import React from 'react' import Error from 'next/error' import fetch from 'isomorphic-fetch' */ } You can specify a name to use for a custom build directory. For example, the following config will create a build folder instead of a .next folder. If no configuration is specified then next will create a .next folder. // next.config.js module.exports = { distDir: 'build' } }) => { //).", "stage-0"] }: any CSS-in-JS solution in your Next app by just including your favorite library as mentioned before in the document. Next.js bundles styled-jsx supporting scoped css. However you can use any CSS preprocessor solution in your Next app by following one of these examples:. This is a known issue with the architecture of Next.js. Until a solution is built into the framework, take a look at this example solution to centralize your routing. Since our first release we've had many example contributions, you can check them out in the examples directory. Please see our contributing.md
https://npm.taobao.org/package/n3tr-next
CC-MAIN-2020-10
refinedweb
423
58.08
Making a Microsoft Excel worksheet to an image file can be quite useful when you might need to use an image of a worksheet in an application or web page. This can be the most effective solution for developers to convert worksheet to image as this solution converts Excel spreadsheets-including all text, formatting and diagrams-to graphic files. Spire.XLS for .NET supports converting Excel worksheets to images. To use this feature, you need to invocate the method - public void SaveToImage(string fileName) in your program or project. You can convert a worksheet to image file(s) with different attributes or options and you can save the image file to disk. Most image formats are supported, such as BMP, PNG, GIF, JPG, JPEG, TIFF, and etc. - Create a new workbook - Load information from excel file and transfer into workbook - Save workbook as any image file format at your will Below is an effective screenshot of the target . png image from Excel spreadsheet. Download Spire.XLS for .NET (or Spire.Office) with .NET Framework 2.0 (or above) together. In your application add Spire.XLS reference. The following code snippet shows how to convert an Excel worksheet to an image in C#, VB.NET. using Spire.Xls; namespace Xls2Image { class Program { static void Main(string[] args) { Workbook workbook = new Workbook(); workbook.LoadFromFile(@"..\..\test.xls"); Worksheet sheet = workbook.Worksheets[0]; sheet.SaveToImage("sample.jpg"); } } } Imports Spire.Xls Namespace Xls2Image Friend Class Program Shared Sub Main(ByVal args() As String) Dim workbook As New Workbook() workbook.LoadFromFile("..\..\test.xls") Dim sheet As Worksheet = workbook.Worksheets(0) sheet.SaveToImage("sample.jpg") End Sub End Class End Namespace
http://www.e-iceblue.com/Tutorials/Spire.XLS/Spire.XLS-Program-Guide/Convert-Excel-to-Image-Worksheet-to-Image-in-C-VB.NET.html
CC-MAIN-2015-22
refinedweb
274
58.48
Learning Resources for Software Engineering Students » Author: Ian Teo, Phang Chun Rong Computer performance can be defined as the rate of work accomplished by a computer system. Even if execution time is not important for a particular application, it may be important to reduce CPU cycles so as to consume less power; from applications running in small battery operated devices to huge data centres. Premature optimization is the root of all evil. I think many people have heard of this quote from Donald Knuth before. This quote is actually misinterpreted frequently, because of the lack of context. Here is the full quote: We should forget about small efficiencies, say about 97% of the time: premature optimization is the root of all evil. Yet we should not pass up our opportunities in that critical 3%. You can find a nice explaination of the quote here. This is not a guide on how to optimize that 3%, but rather, to explain standard optimization techniques that you should apply in all of your code, so that you do not create sub-optimal codes (premature pessimization). If you want to find out more about optimizing that 3%, you can find more at Other Resources below. The 3 techniques you should know are: Using appropriate Data Structures and Algorithms can improve the execution speed of your program trememdously. However, it requires time to learn and understand the nuances of each Data Structure and Algorithm and when to use them. A good example of how Data Structures and Algorithms can improve execution time is with binary search. When finding a particular element in a list, you need to search through the entire list. If it has a million entries, every find will require you to look through 1,000,000 entries. However, if you have a sorted list, you can use binary search to reduce the number of entries you have to look through to around 20! If you are interested in learning more about Data Structures and Algorithms, you can learn more from the following resources: Memory management is important for performance optimization for Computer systems. One of the common techniques in algorithms optimization is space and time trade off, where we increase runtime memory usage to decrease overall runtime. While this theoretically optimizes your system runtime, it might overall slowdown the system due to Thrashing. Thrashing occurs when the system runs out of Random Access Memory and the Operating System swaps main memory to disk memory resulting in significant time spent on disk access. Detecting if the performance slowdown is memory related can be done with appropriate memory profiling. If your system suffers from memory related performance issues, here are some solutions you can adopt to resolve them: Using generators to reduce memory used. Generators are functions that generates a sequence of values. Instead of returning an explicit array upfront, a generator returns a value at each iteration. This can greatly reduce memory usage for large arrays. Sometimes memory usage of your program remains high because the unnecessary variables are yet to be freed from memory. If you are using a garbage collected language like Java, consider tuning your garbage collector to suit your needs. If such options is not good enough, you can explicitly free memory even in garbage collected language. An example from Python is shown below: import gc gc.collect() Using appropriate variable types can also offer memory usage improvement. For example, we should prefer to use primitive int over Integer to reduce the overhead introduce by the Integer Object wrapper. This guide for Java also proposes ways to overcome obstacles introduced by the usage of primitives such as restrictions of JDK collections that requires Object wrappers. Before we can talk about this, we need to know what the computer memory is. Computer memory has different components, registers, L1/L2/L3 cache, RAM, and disk in order of their speed. You can take a look at the extent of the difference of their access speeds here You can also check out this infographic To understand this topic, you only need to know how the cache works. It is okay if you do not understand exactly how the other components, registers, RAM and disk work. The cache is simply a place to store memory so that it can be accessed quickly. The cache is usually split into a few layers, L1, L2 and L3 cache, where the L1 cache is the fastest and L3 is the slowest. The faster a cache is, the more expensive it is. This means that the L1 cache (a few kilobytes big) is smaller than the L2 cache (a few megabytes big) and so on, until the RAM (a few gigabytes big). Whenever data is requested, the computer will first look in the cache for the data. If it exists, this is known as a cache hit. If it does not exist, this is known as a cache miss. When a cache miss happens, a contiguous block of memory containing the requested data is retrieved and stored onto the cache. Because of this, we want to remember these rules to make cache friendly codes: This rule states that recently used memory will likely be used in the near future. This means that making the scope of your variables smaller helps with execution times, as it will likely result in less cache misses. This rule states that memory stored near each other will likely be used in the near future. This means that using contiguous data structures, such as arrays, help improve the execution times. This is because the contiguous block of memory will likely contain the other elements of the array, resulting in less cache misses. An example of a Data Structure that does not do well in this aspect is Linked List. In a Linked List, each node can be stored anywhere on the memory. This means that there will likely be more cache misses when trying to iterate through a Linked List. This can cause Linked Lists to be much slower than what you would expect in theory. This rule is about how multidimensional arrays are stored in memory. Different programming languages have different methods of storing multidimensional arrays. More Information. Using the incorrect method of access can cause many cache misses, resulting in a much slower execution time. Thus, it is important to be aware of which major order the programming language is using. For example, Java uses Row Major Order. We can create a test to see how big an impact using the wrong Major Order can be on the execution time. int size = 10000; int[][] arr = new int[size][size]; int x = 0; //Row Major order accessing long time = System.currentTimeMillis(); for (int i = 0; i < size; i++) { for (int j = 0; j < size; j++) { x += arr[i][j]; } } System.out.println("Row major: " + (System.currentTimeMillis() - time)); //Column Major order accessing time = System.currentTimeMillis(); for (int i = 0; i < size; i++) { for (int j = 0; j < size; j++) { x += arr[j][i]; } } System.out.println("Column major: " +(System.currentTimeMillis() - time)); In the example above, Row major takes around 100ms, while column major takes around 2000ms. You can use the codes above and try it on your own too. This test was done on a typical notebook. Your results may vary based on the hardware of your computer Wikipedia and the University of Maryland have excellent articles which covers everything I have mentioned and more. If you want to know more about Optimization, especially for that critical 3%, these other resources could be useful:
https://se-education.org/learningresources/contents/performance/Performance.html
CC-MAIN-2019-26
refinedweb
1,261
61.97
ttyname(3) BSD Library Functions Manual ttyname(3) NAME isatty, ttyname, ttyslot -- get name of associated terminal (tty) from file descriptor LIBRARY Standard C Library (libc, -lc) SYNOPSIS #include <unistd.h> int isatty(int fildes); char * ttyname(int fildes); int ttyslot(void);ildes refers to a valid terminal type device. The ttyname() function gets the related device name of a file descriptor for which isatty() is true. The ttyslot() function fetches the current process' control terminal num- ber from the ttys(5) file entry. RETURN VALUES The ttyname() function returns the null terminated name if the device is found and isatty() is true; otherwise, a NULL pointer is returned. The ttyslot() function returns the unit number of the device file if found; otherwise, the value zero is returned. FILES /dev/* /etc/ttys SEE ALSO ioctl(2), ttys(5) HISTORY A isatty(), ttyname(), and ttyslot() function appeared in Version 7 AT&T UNIX. BUGS The ttyname() function leaves its result in an internal static object and returns a pointer to that object. Subsequent calls to ttyname() will modify the same object. BSD June 4, 1993 BSD Mac OS X 10.8 - Generated Fri Aug 31 16:25:54 CDT 2012
http://www.manpagez.com/man/3/ttyslot/
CC-MAIN-2018-39
refinedweb
199
62.98
To explain the 140 series of the Federal Information Processing Standards (FIPS) in plain talk and without dozing off is challenging, but let me give it a try: FIPS cryptographic standards specify design and implementation requirements for cryptographic modules. Software and hardware vendors can contract with accredited laboratories to validate their modules against these standards, which then allows others to use those modules in computing environments that must adhere to the U.S. government’s information processing standards. Two questions come to mind. First, what is a cryptographic module? A module might be a piece of software, hardware, or a combination of both. For example, RSAENH.DLL on various Windows platforms is FIPS complaint. Another example would be the Samsung crypto modules running on Android devices like the Galaxy S4. Second question, from a software developer’s perspective, is who must use FIPS compliant modules? The obvious answer is most anyone working directly or indirectly for a department or agency of the United States federal government who is handling sensitive but unclassified data. However, FIPS has also made inroads into private sector healthcare and banking businesses, as both industries store and transmit sensitive data like credit card numbers and personal health information. Some businesses will enforce the use of FIPS compliant algorithms by flipping the “FIPS switch”. This is a setting on operating systems that ensure applications only use FIPS verified cryptography algorithms. You can flip this switch on Windows, as well as OS X and other operating systems and devices. On Windows, the FIPS switch impacts the entire system and all applications, from BitLocker to Internet Explorer and Remote Desktop. The FIPS switch will also impact the code you write. Even if you don’t want to use a FIPS compliant algorithm, if your C# code executes on a Windows machine with the FIPS switch on, you’ll have to use a FIPS compliant algorithm or the code will fail. For example, with C#, the managed .NET implementations of AES encryption are not certified, so the following code fails with an exception. using System.Security.Cryptography; static class Program { static void Main() { var provider = new AesManaged(); } } Unhandled Exception: System.InvalidOperationException: This implementation is not part of the Windows Platform FIPS validated cryptographic algorithms. Instead of using AesManaged, you’ll need to use AesCryptoServiceProvider, which calls into native, verified modules. And just so you know, your application won’t be the only one to face these types of exceptions, as everything from Visual Studio to Internet Explorer and even frameworks like ASP.NET and databases like MongoDB have had (or still have) troubles with FIPS at one point or another. Be careful when you flip on FIPS that you don’t cripple your own development machine, you might need to reserve FIPS for a testing environment. One of the issues you might run into with the FIPS switch is how the system will use brute force to deny access to uncertified encryption modules. There is no awareness from the OS of why an application might choose to use a specific algorithm. In the case of AES the brute force approach might make sense, but when you start to think of algorithms like MD5 the situation is grey. MD5 might be used cryptographically to generate a message digest, but someone might have also chosen MD5 to generate a hash to use as the key value for plain text data in a distributed cache. But, MD5 is considered weak from a crypto perspective, and the following code will also fail with the exception we saw previously, even though the code is using a crypto service provider. using System.Security.Cryptography; static class Program { static void Main() { var provider = new MD5CryptoServiceProvider(); } } To run on a computer with the FIPS switch on, then, you need to be careful about your choice of anything cryptographic. The irony of the FIPS switch is that the system can only prevent what the system knows about. Microsoft programmed the .NET crypto classes to respond with an exception when used inappropriately, but there are other libraries and platforms that will run any kind of cryptographic algorithm on a machine with the FIPS switch on. For example, I can still run the following code in Node on a FIPS enabled Windows machine. var crypto = require("crypto"); var message = "The Magic Words are Squeamish Ossifrage"; var hash = crypto.createHash("md5").update(message).digest("hex"); console.log(hash); I think the question is debatable. The FIPS flag will keep the honest applications honest. But the FIPS flag doesn’t guarantee that an application encrypts the right data, or that an application encrypts data at the right time, or that an application developer doesn’t “work around” the FIPS flag by writing their own algorithm with XOR and clocking out. The FIPS flag also can’t stop a user from storing their passwords on a yellow sticky note affixed to the back of an LCD monitor. The real question should be: “is the data on a FIPS enabled machine more secure than the data on a machine without the FIPS flag?” I’d say the answer is unquestionably a “no”, and system administrators should know the FIPS flag is not a silver bullet for security.
https://odetocode.com/blogs/scott/archive/2014/03/18/working-with-fips-140-crypto-standards.aspx
CC-MAIN-2021-21
refinedweb
872
51.58
I've been getting a lot of PM's asking me to upload my full configs (thanks for the complements btw). I'd prefer to put them on a server somewhere, so I can always link to the latest one, so if anyone wants to loan me some space I'll be grateful. But in the meantime here is the latest config for my (1440x900) laptop. My desktop at work is dual 16x12 so some parts of the configs differ, but I try to keep them as merged as possible. Note: the notes from? … 78#p311378 still applies including the killall issue. If anyone has some ideas on how to avoid it, I'm all ears(eyes?). Note2: I am picky with how I handle multiple displays, so it will behave a bit differently form any standard config. Note3: I switched from iwl3945 to ipw3945 so wlan0 became eth1 New features: blinking "NEW IM" message if pidgin uses the led-notification plug-in from extra/purple-plugin-pack safer key_simulator (sorry if the old caused problems) dzonky: #!/bin/sh FG='#aaaaaa' BG='black' FONT='-*-terminus-*-r-normal-*-*-120-*-*-*-*-iso8859-*' BIGFONT='-*-terminus-*-r-normal-*-*-140-*-*-*-*-iso8859-*' ICONPATH=/home/mstearn/bin/dzen_bitmaps EVENTS="entertitle=scrollend,uncollapse;leavetitle=collapse;button4=scrollup;button5=scrolldown" EVENTSCLEARABLE="entertitle=uncollapse;leavetitle=collapse" WIFI=eth1 logger (){ while true do read line echo "$line" done \ | awk 'BEGIN {print "^fg(white)|^fg()'$1'^fg(white)|^fg()";fflush()}; {print ;fflush()}' \ | sed -re 's/(([^ ]* +){3})/^fg(white)\1^fg()/' #Highlight datetimes } key_symulator(){ key="primer" while [ $key ] do read key xte 'keydown Alt_L' "key $key" 'keyup Alt_L' done } alias my_dzen='dzen2 -xs 1 -h 18 -x $BASE -tw $WIDTH -fg $FG -bg $BG' set_pos (){ # set_pos width_in_pixels BASE=$(echo $BASE + $WIDTH | bc) WIDTH=$1 } BASE=0 WIDTH=0 set_pos 155 watchfile ~/.xmonad-status \ | awk -F '|' '{print $1;fflush()}' \ | my_dzen -e 'button1=menuprint' -ta c -w 160 -m h -l 9 \ | key_symulator & set_pos 240 watchfile ~/.xmonad-status \ | awk -F '|' '/\|/ {print $3 "\n" $3;fflush()}' \ | my_dzen -fn $FONT -e $EVENTSCLEARABLE -ta l -l 1 -u & set_pos 75 gcpubar -h 18 -w 50 -gs 0 -gw 1 -i 2 -s g -l "^i(${ICONPATH}/cpu.xbm)^fg(#222222)" \ | my_dzen -fn $FONT -e '' -ta l& set_pos 60 inotail -f -n 50 /var/log/kernel.log \ | logger KERNEL \ | my_dzen -fn $FONT -e $EVENTS -ta c -l 20 & set_pos 50 inotail -f -n 50 /var/log/messages.log \ | logger MAIN \ | my_dzen -fn $FONT -e $EVENTS -ta c -l 20 & set_pos 60 inotail -f -n 50 /var/log/daemon.log \ | logger DAEMON \ | my_dzen -fn $FONT -e $EVENTS -ta c -l 20 & set_pos 100 while [ $? -ne 1 -a $? -ge 0 ] && echo -n '^fg()' do if (amixer sget Master | grep -qF '[off]') then color='#3C55A6' icon=${ICONPATH}/vol-mute.xbm else color='#7CA655' icon=${ICONPATH}/vol-hi.xbm fi percentage=$(amixer sget Master | sed -ne 's/^.*Front Left: .*\[\([0-9]*\)%\].*$/\1/p') echo $percentage | gdbar -fg $color -bg darkred -h 18 -w 50 -l "^fg(lightblue)^i($icon)^p(2)^fg()" inotifywait -t 30 /dev/snd/controlC0 -qq done \ | my_dzen -fn $FONT -ta l -e \ 'button1=exec:amixer sset Master toggle -q;button4=exec:amixer sset Master 1+ -q;button5=exec:amixer sset Master 1- -q;' & set_pos 45 while echo -n ' ' do ~/bin/dzenWeather/dzen_weather.pl sleep 300 # 5 min done \ | my_dzen -fn $FONT -ta l -w 145 -l 2 -u -e $EVENTSCLEARABLE & set_pos 60 while sleep 1 do line='0' if (exit $(echo `date '+%S'` % 2 | bc)); then #odd seconds line=$(cat /tmp/chat-notify) fi echo $line done \ | awk -F '|' '/0/ {print ""; fflush()}; /1/ {print "^fg(yellow)NEW IM"; fflush()};' \ | my_dzen -fn $FONT -ta c & set_pos 413 CONKY1='^cs() ^tw()^i('$ICONPATH'/temp.xbm)$acpitemp^p(3)${exec if (acpitool -a |grep -qF on-line) ; then echo -n "^fg(green)^i('$ICONPATH'/power-ac.xbm)^fg()"; else echo -n "^fg(red)^i('$ICONPATH'/power-bat.xbm)^fg()"; fi}$battery $battery_time ^fg(white)^i('$ICONPATH'/net-wifi.xbm)(${wireless_essid '$WIFI'} ${wireless_link_qual_perc '$WIFI'} ^fg(#BA9093)${addr '$WIFI'}^fg(#80AA83)^p(3)^i('${ICONPATH}'/arr_up.xbm)${upspeedf '$WIFI'}^p(3)^i('${ICONPATH}'/arr_down.xbm)${downspeedf '$WIFI'}^fg(white)) LOADS: ^fg(#ff0000)${loadavg 1 2 3} UPTIME: $uptime fs ' conky -t "$CONKY1" \ | my_dzen -fn $FONT -e $EVENTSCLEARABLE -ta r -l 10 -w 300& set_pos 182 CONKY2='^tw()${time %a %b %d ^fg(white)%I:%M^fg():%S%P}^p(5) ${execi 3600 (cal -3 | awk "BEGIN {print \\"^cs()\\"}; {print \\"^p(15)\\", \\$0}")}' conky -t "$CONKY2" \ | my_dzen -fn $BIGFONT -e $EVENTSCLEARABLE -ta r -w 550 -l 8 & # vim:set nospell ts=4 sts=4 sw=4: xmonad.hs: {-# OPTIONS_GHC -fglasgow-exts #-} -- -- xmonad example config file. -- -- A template showing all available configuration hooks, -- and how to override the defaults in your own xmonad.hs conf file. -- -- Normally, you'd only override those defaults you care about. -- import XMonad import System.Exit import XMonad.Actions.CopyWindow import XMonad.Actions.DwmPromote import XMonad.Actions.WindowBringer import XMonad.Hooks.EwmhDesktops import XMonad.Hooks.DynamicLog import XMonad.Hooks.UrgencyHook import XMonad.Layout.Tabbed import XMonad.Layout.Magnifier import XMonad.Layout.Maximize import XMonad.Layout.NoBorders import XMonad.Layout.ResizableTile import XMonad.Prompt import XMonad.Prompt.Shell import qualified XMonad.StackSet as W import qualified Data.Map as M import qualified Data.Set as S import qualified Data.List as L (deleteBy,find,splitAt,filter,nub) import Control.Monad (when) -- The preferred terminal program, which is used in a binding below and by -- certain contrib modules. -- myTerminal = "roxterm" -- Width of the window border in pixels. -- myBorderWidth =1Mask -- The mask for the numlock key. Numlock status is "masked" from the -- current modifier status, so the keybindings will work with numlock on or -- off. You may need to change this on some systems. -- -- You can find the numlock modifier by running "xmodmap" and looking for a -- modifier with Num_Lock bound to it: -- -- > $ xmodmap | grep Num -- > mod2 Num_Lock (0x4d) -- -- Set numlockMask = 0 if you don't have a numlock key, or want to treat -- numlock status separately. -- myNumlockMask = mod = "#222233" myFocusedBorderColor = "#ff0000" -- Default offset of drawable screen boundaries from each physical -- screen. Anything non-zero here will leave a gap of that many pixels -- on the given edge, on the that screen. A useful gap at top of screen -- for a menu bar (e.g. 15) -- -- An example, to set a top gap on monitor 1, and a gap on the bottom of -- monitor 2, you'd use a list of geometries like so: -- -- > defaultGaps = [(18,0,0,0),(0,18,0,0)] -- 2 gaps on 2 monitors -- -- Fields are: top, bottom, left, right. -- myDefaultGaps = [(18,0,0,0)] ------------------------------------------------------------------------ -- Key bindings. Add, modify or remove key bindings here. -- myKeys conf@(XConfig {XMonad.modMask = modMask}) = M.fromList $ -- launch a terminal [ ((modMask .|. shiftMask, xK_Return), spawn $ XMonad.terminal conf) -- launch dmenu , ((modMask, xK_p ), spawn "exe=`dmenu_path | dmenu` && eval \"exec $exe\"") -- mod-shift-g goto window menu -- mod-shift-g bring me window menu , ((modMask, xK_g ), gotoMenu) , ((modMask .|. shiftMask, xK_b ), bringMenu) -- mod-shift-d (re)launch dzen , ((modMask .|. shiftMask, xK_d ), spawn "killall dzonky dzen2 inotail inotifywatch watchfile ; sleep 1; exec /home/mstearn/bin/dzonky") , ((modMask .|. shiftMask .|. controlMask, xK_d ), spawn "dzonky") , ((modMask .|. controlMask, xK_Escape ), spawn "xkill") -- mod-\ toggle maximizedness , ((modMask, xK_backslash), withFocused (sendMessage . maximizeRestore)) , ((modMask, xK_x), shellPrompt $ defaultXPConfig ) -- launch gmrun , ((modMask .|. shiftMask, xK_p ), spawn "gmrun") -- amarok , ((0, 0x1008ff14 ), spawn "amarok -t") , ((shiftMask, 0x1008ff1d ), spawn "amarok -r") , ((0, 0x1008ff1d ), spawn "amarok -f") -- reset modifiers if i hit Web/Home , ((0, 0x1008ff18 ), spawn "xmodmap ~/.xmodmap") -- close focused window -- , ((modMask .|. shiftMask, xK_c ), kill) , ((modMask .|. shiftMask, xK_c ), kill1) --_j ), windows W.focusDown) , ((modMask, xK_Tab ), windows W.focusDown) -- Move focus to the previous window , ((modMask, xK_k ), windows W.focusUp) , ((modMask .|. shiftMask, xK_Tab ), windows W.focusUp) -- Move focus to the master window , ((modMask, xK_m ), windows W.focusMaster ) -- Swap the focused window and the master window , ((modMask, xK_Return), dwmpromote) -- (Just "xmonad") True) ] ++ -- mod-[1..9] @@ Switch to workspace N -- mod-shift-[1..9] @@ Move client to workspace N -- mod-control-shift-[1..9] @@ Copy client to workspace N [((m .|. modMask, k), windows $ f i) | (i, k) <- zip (XMonad.workspaces conf) [xK_1 .. xK_9] , (f, m) <- [(ungreedyView, 0), (W.shift, shiftMask), (copy, controlMask .|. shiftMask)]] -- -- mod-{w,e,r}, Switch to physical/Xinerama screens 1, 2, or 3 -- mod-shift-{w,e,r}, Move client to screen 1, 2, or 3 -- ++ [((m .|. modMask, key), screenWorkspace sc >>= flip whenJust (windows . f)) | (key, sc) <- take (length myDefaultGaps - 1) $ zip [xK_w, xK_e] [0..] , (f, m) <- [(W.view, 0), (W.shift, shiftMask), (W.greedyView, controlMask)]] -- like view, but do nothing if is visible ungreedyView i s = if i `elem` (map (W.tag . W.workspace) $ W.visible s) then s else W.view i s ------------------------------------------------------------------------ -- Mouse bindings: default actions bound to mouse events --) -- resize the windows in tiled mode , ((modMask, button4), (\w -> focus w >> sendMessage MirrorExpand)) , ((modMask, button5), (\w -> focus w >> sendMessage MirrorShrink)) ] ------------------------------------------------------------------------ -- = maximize (tiled ||| Mirror tiled ||| noBorders (tabbed shrinkText defaultTConf) ) where -- default tiling algorithm partitions the screen into two panes tiled = ResizableTall =? "pidgin" --> doF (W.shift "9") , resource =? "desktop_window" --> doIgnore , resource =? "kdesktop" --> doIgnore , resource =? "kicker" --> doIgnore , resource =? "stalonetray" --> doIgnore , title =? "panel" --> doIgnore ] ------------------------------------------------------------------------ -- Status bars and logging -- Perform an arbitrary action on each internal state change or X event. -- See the 'DynamicLog' extension for examples. -- -- To emulate dwm's status bar -- -- > logHook = dynamicLogDzen -- redbeardPP = defaultPP { ppHiddenNoWindows = dzenColor "#33FF00" "" , ppHidden = dzenColor "white" "" , ppCurrent = dzenColor "yellow" "" , ppVisible = dzenColor "#FF2222" "" --, ppUrgent = dzenColor "red" "yellow" , ppSep = " | " , ppWsSep = "\n" , ppTitle = id , ppLayout = \s -> " " , ppOutput = writeFile "/home/mstearn/.xmonad-status" . ("^cs()\n"++) } myLogHook = do ewmhDesktopsLogHook dynamicLogWithPP redbeardPP return () myUrgencyHook = withUrgencyHook dzenUrgencyHook { args = ["-bg", "yellow", "-fg", "black" ,"-p", "3", "-xs", "1"] } ------------------------------------------------------------------------ -- Now run xmonad with all the defaults we set up. -- Run xmonad with the settings you specify. No need to modify this. -- main = xmonad $ myUrgencyHook defaults -- A structure containing your configuration settings, overriding -- fields in the default config. Any you don't override, will -- use the defaults defined in xmonad/XMonad/Config.hs -- -- No need to modify this. -- defaults = defaultConfig { -- simple stuff terminal = myTerminal, borderWidth = myBorderWidth, modMask = myModMask, numlockMask = myNumlockMask, workspaces = myWorkspaces, normalBorderColor = myNormalBorderColor, focusedBorderColor = myFocusedBorderColor, defaultGaps = myDefaultGaps, -- key bindings keys = myKeys, mouseBindings = myMouseBindings, -- hooks, layouts layoutHook = myLayout, manageHook = myManageHook, logHook = myLogHook } -- vim:nospell: Offline thanks very much redbeard... also see your pm's gonna try to get this working tomorrow Offline Hey guys, I am just new to xmonad and dzen2, and I installed both. I would like to get them going but don't quite know the right starting point and browsing this thread has given some great inspiration for possibilities! How would I make a log-in session in GDM for xmonad + dzen2? And what are some basic/important shortcuts? I'd like to spend some of this weekend setting it up Offline Welcome colbert! create ~/.xmonad and ~/bin directories and add ~/bin to your $PATH. Put xmonad.hs into .xmonad and dzonky into bin. Many dzen scripts (including mine) require the svn version of dzen and the dzen gadgets. The default key bindings are listed in man xmonad. dzen doesn't have a place for tray applications, so if you need them I'd suggest running stalonetray. The default xmonad config: … _Config.hs The guide for *dm: … .2C_gdm.29 My configs: PS- thanks for the server space Evanlec Offline You're very welcome! And glad that it also makes it easier for everyone to access these great configs... Now finally to get mine setup the way I REALLY want it btw, redbeard has your setup been added to the xmonad/dzen setups page? if not, it should be. Offline I can't get dzen to show anything, I put that dzonky in /home/user/bin. Is there more to it ? Offline I can't get dzen to show anything, I put that dzonky in /home/user/bin. Is there more to it ? dzonky is just a script that executes conky-cli and dzen. If you don't already have them, you'll need to install both dzen and conky-cli (from AUR) and then execute the dzonky script, either in your .xinitrc or however you like. Last edited by thayer (2008-01-19 17:21:51) thayer williams ~ cinderwick.ca Offline Okay I grabbed conky-cli from AUR and did makepkg fine, but on -U I get: bobby@dabox:~/installs/conky-cli$ pacaur conky-cli-1.4.9-2-i686.pkg.tar.gz loading package data... done. checking dependencies... error: replacing packages with -A and -U is not supported yet error: you can replace packages manually using -Rd and -U error: failed to prepare transaction (conflicting dependencies) :: conky-cli: conflicts with conky Do I remove conky first, or? Offline Just writing to inform you about a new bit that you can use with dzen: dmplex a simple multiplexer. dmplex is useful if you want to combine multiple input sources into a single dzen instance. Basically we split the dzen window up into numbered sections and let dmplex handle the layout of the sections, you just supply the value of the section you want to alter. Usage example: # First we create a named pipe $ mkfifo dmpipe # Listen on the pipe and multiplex the input to dzen $ tail -f dmpipe | dmplex | dzen2 # Now we can asynchronously write to the pipe and have dmplex handle the layout for us $ echo "2 second section " > dmpipe $ echo "1 first section ^r(100x5)" > dmpipe $ echo "29 29th section" > dmpipe $ echo "1 first section changed" > dmpipe The syntax is: section number<space>actual dzen input You get the trick.. Have fun, Rob. Last edited by gotmor (2008-01-22 14:03:26) Offline Will this be included in dzen2 gadgets or should I make a package for it? I don't see it in SVN so I'll package it. Last edited by Xilon (2008-01-23 06:10:07) Offline Will this be included in dzen2 gadgets or should I make a package for it? I don't see it in SVN so I'll package it. Yes, it will be included in dzen2. I just wanted to hear some oppions/suggestions before the inclusion, so that I can make sure everything works fine. Offline The multiplexer is a very good idea. As a way of learning Python, I'm writing a multi-threaded script that prints dzen strings --if I needed it, I guess other people will need that functionality too. A good thing about this program is that it can be expanded to manage how the different sections are placed / spaced (instead of having to guess-and-check ^p(), users could e.g. tell dmplex to space out the sections evenly). ...hmm I could even write that part myself! Keep up the good work. Edit: after thinking about this, I see that "automatic spacing" is not feasible. To get proper spacing, the input would have to be interpreted, and that requires dzen. So maybe the multiplexer should be integrated into the main dzen2 program to allow these features. Edit2: what if a process is not finished writing to the pipe, and another process begins to write to it? Last edited by peets (2008-01-26 18:51:46) Offline Building complex widgets with dzen: Further details on the Wiki. That *IS* nice! I've been messing around with dzen again for the past two days and actually I had a thought if something like that was possible, but I guess the thought came in reality just now! The actual thought was something like "a dzen based mpd client" or something alike. I was about to ask something but can't remember it anymore, your sample just messed up my mind ..btw.. still working on those 8x8 xbm icons. I reinstalled arch on my new (or.. "new" as in "new to me") laptop and there's been a lot of other stuff I've had to do first, but now I' back on track. Actually this brings up one more question: What icons should there at least be in a dzen icon set? Maybe if there was an "universal naming list" of some sort to all the common icons, that'd also help much, both the users and the one's who are drawing or converting icons from already available sets. Any ideas, suggestions, thoughts on that one? And now I remember the original question also: From what I read in these post, there are good, better and different ways to fetch info to parse to dzen. What do you use.. or let me make this clearer.. what would you make your alias look like if it was to fetch info on XYZ? An example (I just came up with) to get cpu usage: vmstat | tail -1 | awk '{print 100-$15}' Sakari Offline @sm4tik Just wanted to say, looking forward to your icon package Cthulhu For President! Those fanspeed icons are fantastic! Too bad I don't use fanspeeds in my monitors At the moment, here's how they look on my machine: At first I was a bit doubtful at the tiny (8x8) size, but they look magnificent. Cthulhu For President! Offline Those fanspeed icons are fantastic! Too bad I don't use fanspeeds in my monitors At the moment, here's how they look on my machine: At first I was a bit doubtful at the tiny (8x8) size, but they look magnificent. Heh.. it's FileSystem really, but now that you mentioned it, I'll have to come up with something totally different for that one. The filesystem icon was and is actually the one with most headache. I don't know what it's supposed to look like (in 8x8)?! So, back to drawing board. 8x8 is actually a lot.. hmm.. it's (8x8)^2.. so.. the possibilities are ..almost infinite Anyway, glad you ( < buttons ) like them and I'm very glad to see a shot like that, and also if you (and that is all of you!) have any improvements you'd like to propose, please read the README or use the email form of this forum to contact me directly. Thanks, Sakari ps. I'm not using xmonad myself, so I _must_ be missing an icon or two.. EDIT: @ buttons Thanks for inspiration! I think I'll add a fanspeed monitor to my upcoming bar just because of you (though I can hear the speed without a monitor.. ) Last edited by sm4tik (2008-02-01 22:32:34) Offline Hey guys, can I have a little bit of help ? I dont understand how the menuexec thing is working. I am trying to do a wallpaper change script, not like change my wallpaper a lot, just thought it may be useful for someone. # background changer script . ~/.dzen2/dzen.conf # direcory with images DIR='/home/tch/images' COUNT=`ls -l $DIR | wc -l` COUNT=`expr $COUNT - 1` # commad to set the background file for i in $(ls $DIR); do echo $i >> file done dzen2 -fn $FONT -x 800 -y 0 -tw 30 -ta c -sa l -w 400 -p -l $COUNT -m < file -e "entertitle=uncollapse;leaveslave=collapse" Can someone explain how to execute a command ( SETBG ) with the selected menu option as an argument ? I hope it is clear. Offline I'm working on a "dzonata" script for mpd and I'm almost finished, but one thing I'm missing is a way to convert seconds into minutes. echo "obase=60;${current_seconds}/1" | bc is the closest one so far, but if current position is less than a minute, I haven't found a way to make the "0 minutes" show up. I know I could use mpc's output with sed to get the "position:total", but I was inspired by bashmp and found netcat (which bashmp is "based" on) to be a very useful tool when messing around with mpd. Offline How do I get different sleep timers for different things? I want a 1 second timer on my eth0 monitor, and a 60 second timer on my Gmail notifier. Offline @ dax965 quarks has a nice setup nice setup which does just that. Last edited by sm4tik (2008-02-03 12:10:51) Offline How do I get different sleep timers for different things? I want a 1 second timer on my eth0 monitor, and a 60 second timer on my Gmail notifier. dzen is scriptable in any language, so there are many many ways this can be done. Probably the best way is to use two instances of dzen, each getting updated at different intervals. Offline
https://bbs.archlinux.org/viewtopic.php?id=40637&p=5
CC-MAIN-2016-30
refinedweb
3,392
65.52
How to Add Hosts in an Active Directory Domain Updated: January 22, 2011 Applies To: Virtual Machine Manager 2008, Virtual Machine Manager 2008 R2, Virtual Machine Manager 2008 R2 SP1 You can use the Add Hosts Wizard to add one or more virtual machine hosts in an Active Directory Domain Services (AD DS) domain. The domain can be either a domain that has a two-way trust with the domain that VMM server is in or a domain that does not have a two-way trust with the Virtual Machine Manager (VMM) server’s domain. You can also use the Add Hosts Wizard to add a host that is on a perimeter network, after you have first installed a VMM agent locally on the host. For more information, see How to Add Hosts on a Perimeter Network. When you add a Windows-based host, VMM automatically installs or enables the appropriate version, VMM automatically installs the correct version of Virtual Server 2005 R2 if it is not installed already. To add virtual machine hosts in a domain In any view in the VMM Administrator Console, in the Actions pane, click Add hosts to open the Add Hosts Wizard. On the Select Host Location page, do the following: - Click Windows Server-based host on an Active Directory domain, and then do one of the following: - To add hosts on a trusted domain, type the credentials for a domain account with administrative rights on all hosts you want to add. If you are adding a host cluster, the account must have administrative rights on all nodes of the cluster. You cannot use the same domain account that is used as the VMM service account to add or remove a Hyper-V or Virtual Server host from VMM. For more information, see Hardening Virtual Machine Hosts Managed by VMM (). - To add hosts that are not in a domain that has a two-way trust with the VMM server’s domain, clear the Host is in a trusted domain check box, and then type the credentials for a local account with administrative rights on all hosts you want to add. - Click Next. On the Select Host Servers page, in the Domain box, type the domain name or domain alias of the hosts you want to add, and then do one of the following: - In the Computer name box, type the NetBIOS name, the IP address, or the fully qualified domain name (FQDN) of a host in the specified domain. To add a host in a different domain, change the value in the Domain box, and then enter the NetBIOS name. If the host is in a disjointed namespace, use the host’s fully qualified domain name (FQDN) and select the Skip the Active Directory name verification check box. - If the specified domain has a two-way trust with the VMM server’s domain, you can click Search to open the Computer Search dialog box, and then search for computers you want to add as hosts. For more information, see How to Search for Hosts (). On the Configuration Settings page, do the following: - In the Host group list, select a host group that will contain the hosts or host cluster, or accept the default host group, All Hosts, which is the parent host group of all hosts and host groups. one or more of the computers you are adding is a host or a library server that is currently being managed by another VMM server, select the Reassociate host with this Virtual Machine Manager server check box to associate those hosts with the current VMM server. On the Host Properties page, do the following: - In the Add the following path box, specify a virtual machine default path for storing virtual machines deployed on the hosts, and then click Add. - In the Remote connection area, by default, the Enable remote connections to virtual machines on these hosts check box is enabled and set to use the global default port setting. To disable remote connections, clear the check box. To use a different port for remote connections, enter a value from 1–65,535 in the Remote connection port box. To change the global default port setting, see How to Configure Remote Access to Virtual Machines (). On the Summary page, click Add Hosts.
https://technet.microsoft.com/en-us/library/cc917923.aspx
CC-MAIN-2018-39
refinedweb
716
58.66
StructureStructure The OpenMW project is made up of three distinct subsystems: AppsApps openmw- the actual game openmw-cs- replacement for the Morrowind Construction Set openmw-essimporter- savegame importer openmw-iniimporter- morrowind.ini importer openmw-launcher- the game launcher GUI openmw-wizard- installation wizard for the game files required to run OpenMW. esmtool- esm/esp analysis tool bsatool- bsa analysis tool ComponentsComponents A collection of reusable components. Most of these components are used by more than one application within the OpenMW project. ExternExtern The Extern subsystem consists of libraries that are not part of the OpenMW project, but of which a copy is kept in the repository anyway: oicsoics Fork of the OIS Input Control System project, a library that makes it easy to handle input configurations for mouse, keyboard and joysticks by routing bindings through abstract channels. We use it for OpenMW's in-game key binding menu. When OpenMW switched to SDL2 input, modifications to oics were made to use SDL2 for input in place of OIS. The abstract design of oics made it relatively painless for OpenMW to start supporting joysticks. osg-ffmpeg-videoplayerosg-ffmpeg-videoplayer A port of the ogre-ffmpeg-videoplayer project, which is now in extern/ for historical reasons. We might move it to components/ at some point. osgQtosgQt A small glue utility between osg and Qt. World ModelWorld Model TerminologyTerminology The world-model terminology is confusing. We have - the original terminology (MW), which is rather unintuitive and partially overlaps with the C++ language terminology, which creates additional ambiguity. - the new terminology (NEW), which is reasonable, but incompatible with MM and overlaps even stronger with the C++ language terminology, which creates even more additional ambiguity. The current state of OpenMW (code and comments) is an unhealthy mix between these two. ID (MW): A certain type of item, which is internally represented as a record. Each ID is identified by a unique string. IDs are immutable. Example: katana_goldbrand_unique Reference (MW): An instance of an ID, that exists either somewhere in the world or in a container/inventory. A reference consists of the name of its ID and some additional data. References are mutable. Not to be confused with C++ references. Example: The sword Goldbrandthat you have in your inventory after finishing the Boethiahquest in Morrowind. Class (NEW): New name for IDs (MW). Not to be confused with C++ classes. We don't have a C++ class per Class (NEW). We have however a C++ class per type of Class (NEW), e.g. Weapon, Misc, Static, … Object (NEW): New name for References (MW). Not to be confused with C++ objects. Instance (NEW): Synonym for Object Record StoreRecord Store The world model is split in two parts: An immutable part and a mutable part. The record store represents the immutable part. The esm/esp file format contains a list of records, where each record is composed of a type ID and additional data. Our world model does mirror this structure. Individual records can by identified by a string ID, a numeric ID, a coordinate-pair or not at all (depending on the type). For each type of record we also have a list of these records, instantiating the MWWorld::Store template. There are some variations depending on the record type. The class implementing the record store is named ESMStore (in the MWWorld namespace). The ESMStore merges records from multiple ESM/ESP files into a single store. The merging implementation can vary depending on the record type. A const reference to the current ESMStore can be obtained from the `MWBase::World class instance. Dynamic RecordsDynamic Records Some types of records can be generated dynamically during gameplay and then accessed via the record store like any other record. Examples: - custom classes - custom enchantments - potions created by alchemy -modifications to levelled lists via the console CellCell A cell record consists of some header-type data fields and a collection of references. Of these only the header data is stored in the record store. The collection of references is mutable and as such needs to be treated separately. InfoInfo For every practical purpose each info record is a sub-records of a dialogue record. The OpenMW info storage implementation reflects this structure. The esm/esp format does not. CellsCells The World of an OpenMW game is composed of cells. Each cell consists of an immutable part (stored in the record store) and a list of objects, which make up the mutable part of the world model. TerminologyTerminology -Inter neighboring cells). In this case no interior cell can be active. ObjectsObjects The mutable part of the world model is made up from objects (called references in the original MW terminology). An object can either be located in a cell or in a container (which is also located in a cell). Each cell and each container keeps a separate list for each record type that can be referenced. Each list is made up from a sequence of ESMS::LiveCellRef objects (specialized for the respective record type). Please note that the LiveCellRef only contains the persistent state. State only valid while the object is in an active cell is stored elsewhere. There is no polymorphism used here, which is a design flaw that originated at a very early development stage and could not be fixed anymore without a total rewrite, once everyone agreed that it was a flaw. Various workarounds are in place now. Unified Object AccessUnified Object Access Unified access to objects is possible through the MWWorld::Ptr class. There are several ways to obtain a Ptr: - directly from an object in a cell or a container - via the record ID (all referencable records have string IDs), if there is only one object of this record -for NPCs and Creatures, via a CreatureStats::ActorId MWWorld::Class Another workaround is the MWWorld::Class hierarchy. Through the static function MWWorld::Class::get from a Ptr a suitable instance of a derived Class-class can be obtained that polymorphically implements record type specific behaviour. PlayerPlayer The player is a special object that references a record of the type NPC. The player is by definition always in an active cell. The player object is not stored within the list of objects of that cell. Instead it is handled separately by the World class. This can be ignored most of the time though. The player can be access like any other object via its ID "player". The world class also provides a function that returns the player's MWWorld::Ptr. Please note that the esm/esp files don't contain a player object. Instead this object is inserted when the player begins a new game. Rendering ArchitectureRendering Architecture Our rendering code is based on OpenSceneGraph, which is based on OpenGL. The best way to learn about OpenSceneGraph is to read its code and mailing list/forum, the mailing list/forum has many helpful explainers when you search for a particular topic. Because OpenSceneGraph is a very powerful framework, and as they say "with great power comes great responsibility" - it is best to decide on some common usage practices, in order to avoid any problems. Scene graph basicsScene graph basics The scene graph is made up of: StateSet: a collection of rendering states (e.g. Textures). Node: a node in the scene graph. A Node can have a StateSet. Group: a Node that has children Nodes. Drawable: a Node that contains drawing commands. The simplest implementation of a Drawable is the Geometry class, which consists of vertices, triangles and texture coordinates. All of these classes are polymorphic and can also be customized by way of callbacks. Please note some OSG examples still make use of a Geode class to attach Drawables, this is obsolete now as Drawables can be placed directly as a node in the scene graph. Geode should not be used. In order to inspect or manipulate a scene graph, you can use a NodeVisitor. Node masks are an important tool to control which parts of the scene graph are used for which purpose. The frame loopThe frame loop The rendering portion of the frame loop is structured into several phases. You can see these in action with the profiling panel brought up by the F3 key in-game. Threading considerationsThreading considerations The default threading model in OpenSceneGraph, DrawThreadPerContext, requires some care when dealing with modifications to rendering data ( Drawables and StateSets). Any such data that is currently in (or has been in) the scene graph (referred to as 'live' data) must be assumed as unsafe to modify because it could be in use by the drawing thread. What is not allowed: - Modify a live StateSet(e.g. by adding/removing StateAttributesor modifying a contained StateAttribute) - Modify a live Drawablein a way that affects its drawing implementation What is allowed: - Remove a Drawablefrom, or add it to the scene graph (the scene graph itself is not part of rendering data) - (un)assign a StateSetfrom a node - the StateSetsused for a rendering traversal are stored separately, so this is fine. - Make harmless modifications that don't affect the renderer, like adding callbacks, changing names, etc. Still be very careful, though. - Whenever you are dealing with a bug that you think may be threading-related, you can use the OSG_THREADING=SingleThreadedenvironment variable to test your theory. As long as you stick to the above rules, though, you shouldn't run into any issues. To make actual modifications, one of the following techniques may be used: - Clone the original object and modify that copy. This is perfect for infrequent modifications when we don't care about the performance overhead of object cloning/deletion. Example - Use a double buffering technique to manage access. We have a StateSetUpdater class that implements this technique for StateSets. Example - Rather than changing the object itself, inject the change on the fly where it's needed, e.g. into the CullVisitorby using a CullCallback. Example - Set the object's dataVarianceto DYNAMIC, so that the draw traversal knows that it has to synchronize that object. Never do this in performance critical areas, or at all, really. Just one DYNAMICobject will make the threading useless and probably halve your frame rate. Important componentsImportant components SceneManagerSceneManager The SceneManager is where game object's rendering nodes (.nif or .osg) are loaded, prepared, optimized and instanced, then stored in cache for future use. Instancing refers to making a copy of the object that is safe to use and modify, while the original 'template' object remains read-only. A state sharing mechanism is employed that will share StateSets across different objects in order to improve rendering efficiency. This means that for any object loaded through the SceneManager it is not safe to modify its StateSets because they could also be in use by another object. It is advised to create a copy of the StateSet before making a modification (just as discussed in Threading considerations). The StateSetUpdater already does this for you by creating a copy of the StateSet on the first frame its used. OptimizerOptimizer The optimizer's job is to restructure a scenegraph in a way that is functionally equivalent yet increases the rendering speed. It does this for example by merging Geometries with the same StateSet into one Geometry, removing redundant nodes, and so on. The Optimizer is run automatically on any object loaded through the SceneManager. It has been known to happen that the optimizer contained bugs, which are then (usually) promptly fixed. These bugs could cause objects to render incorrectly, cause incorrect collisions or spam warnings on the console. In order to check if a bug is related to the Optimizer, we can run OpenMW with the OPENMW_OPTIMIZE=OFF environment variable. If the bug disappears with that flag, we can further narrow down which specific optimization may be causing it, by disabling them in order: OPENMW_OPTIMIZE="~MERGE_GEOMETRY" OPENMW_OPTIMIZE="~MERGE_GEOMETRY|~REMOVE_REDUNDANT_NODES" # If the bug does not appear now, the first optimization (FLATTEN_STATIC_TRANSFORMS) has to be the culprit. UtilitiesUtilities Aside from the optimizer, the sceneutil/ directory contains many more OpenSceneGraph "extensions" made for OpenMW purposes, for example our lighting system. It is a good idea to be familiar with these. There is also a scene graph debugger (invoked in-game with the "showSceneGraph" (ssg) console command), this will print out the scene graph representation of any given object, or even the whole scene (if no object is given). The results are printed to a text file. GUI ArchitectureGUI Architecture OpenMW's game UI is built on MyGUI. To learn about MyGUI visit the code documentation, its wiki, the source code and more specifically the examples contained in the source code. This article will explain MyGUI usage specific to OpenMW. Layout systemLayout system MyGUI's XML format for layouts/skins has been used for the majority of the OpenMW GUI. The resulting files can be found here. Currently the formats are lacking a complete documentation, but the amount of examples that are available should more or less make up for this. MyGUI includes a Layout Editor and Skin Editor application, but unfortunately these tools are difficult to use for us due to specifics of the Morrowind GUI resources, e.g. BSA archives and case insensitivity. In the past we made a MyGUI plugin to work around this, which can still be found in the 'plugins' directory of the OpenMW code. However, this plugin is currently not working as it has not been updated after OpenMW switched to OpenSceneGraph rendering. So for the time being, XML files have to be edited by hand (which, admittedly, is more convenient in many cases anyway). The coordinates in MyGUI layouts (left, top, width, height - always in that order) are given in pixels and relative to the parent widget. The GUI system internally works in pixels, even if GUI Scaling is enabled. The GUI scaling option works by pretending to the GUI that the screen is smaller than it really is, and then using the graphics card to do the scaling. Widgets also have an Align property. The Align does not have any effect on the initial placement of the widget, rather, it controls what will happen when the parent widget changes size. For example, if the parent widget's width increases by 1 pixel, and our widget is set to 'HStretch', its width will also increase by 1 pixel. If in that same example our widget instead uses 'Right' then the widget will move to the right by 1 pixel. Here is the list of possible alignments, which can sometimes be combined for example 'Left Bottom'. Please note that most of OpenMW's skin files were written using a deprecated syntax. The now preferred and more powerful syntax is the ResourceLayout, for which an example can be found here. This syntax more closely resembles the one used for layout files. Apart from using skin/layout files it is still required to use C++ code in order to connect widget events to certain actions. The code for this can be found in the mwgui directory. Box layoutingBox layouting Because the above mentioned pixel-based layouting system can be a bit too limited and cumbersome for some cases, we have created a new box-layout system for OpenMW's purposes that functions similar to the way traditional GUI toolkits like Qt or GTK work. It is usually preferred to use this system for new layouts instead of positioning widgets by hand. The new system introduces the following widgets: HBox, VBoxHBox, VBox Automatically positions and resizes their child widgets, horizontally or vertically. Available user strings are 'Spacing' (empty space between each widget), 'Padding' (outer padding), and 'AutoResize'. If 'AutoResize' is true, the box resizes itself to the requested size of the child widgets. Otherwise, the box will stay at its given size. Child widgets can set the user string 'HStretch' (for HBox) or 'VStretch' (for VBox). If stretching is enabled, the widget will fill up all the space available to it. Otherwise, the widget will remain in its original size i.e. the box will only control that widget's position. Boxes can be nested within each other e.g. to produce a table layout. AutoSizedButtonAutoSizedButton This widget will resize itself to fit its label. Text padding defaults to (24, 8) but can be overriden by the 'TextPadding' user string. If the 'ExpandDirection' user string is set to 'Left' then the button will accordingly move to the left upon expanding its size (or to the right on contracting). This is only relevant if the button is not part of a Box. AutoSizedEditBoxAutoSizedEditBox Like AutoSizedButton, but for multi-line text. The widget's width is static, but the height will be adjusted based on the number of lines required. SpacerSpacer An empty widget for use with HBox/VBox that fills all available space. Spacers can be used to set the alignment of sibling widgets. For example, a box containing a spacer, a button and another spacer (in that order) results in a centered button. Escape sequencesEscape sequences As the MyGUI wiki explains, text widgets support escape sequences starting with '#', for example: #ff0000textresults in a red text. #{foobar}will be replaced with the value for foobarin the MyGUI::LanguageManager. For OpenMW in particular, we have used a callback into the MyGUI::LanguageManager to implement the following meanings: #{setting=<section,setting>}will be replaced with the relevant setting value in the user's OpenMW settings.cfg. #{sCell=<cell id>}will be replaced with the translated name for a Morrowind cell (this is usually the same as the Cell ID, but the Russian edition of Morrowind uses a separate translation table) #{fontcolour=<name>}will be replaced with the color of FontColor_color_<name>in openmw.cfg(imported from Morrowind.ini). The output format is R,G,B with RGB ranging from 0 to 1. #{fontcolourhtml=<name>}is the same as above, except that the output format is a #followed by an HTML color code. #{<gamesetting>}will be replaced with the value of the Game Setting <gamesetting>in the Morrowind Data Files. Leveraging this replacement mechanism at runtime requires one to use the setCaptionWithReplacing method rather than setCaption. Colour-codes (like #ff0000) are always used regardless of whether setCaption or setCaptionWithReplacing is used. To use escape sequences in a layout one must use the 'Caption' property. To avoid # characters in a caption being treated as an escape sequence, this character needs to be escaped by adding another # character: ## produces #. MyGUI provides the function MyGUI::TextIterator::toTagsString to do this. When retrieving text from a user-filled text box, you may want to do the opposite by calling getOnlyText(). Escape sequences are also supported in skins. In order for this to work in legacy skins, the version must be set to 1.1. Example
https://gitlab.com/OpenMW/openmw/-/wikis/development/architecture
CC-MAIN-2021-25
refinedweb
3,119
54.73
JavaScript –). The introduction didn't paint a flattering picture of the JavaScript language, but the truth is JavaScript is a good language. The biggest sources of pain when programming with JavaScript aren't because of the language. The biggest pains come from: Most of the tools we use in a Visual Studio environment are geared to languages targeting the CLR. If you program in C# or Visual Basic, you'll be assisted by Intellisense, class browsers, class diagrams, code snippets, code analysis, and a world class debugger. The menus will list refactoring commands, and unit testing is only a few keystrokes away. Contrast the above with the experience of programming with JavaScript. There is no Intellisense (although the feature is coming in the next version of Visual Studio), the debugger is finicky, and most of the other tools listed above are missing entirely. Of course, the majority of the code we write in Visual Studio is not JavaScript, but as the demands of the web have required more scripting, we've started to need better tools. The lack of tools makes JavaScript more difficult to work with. JavaScript is hosted by many different types of web browsers, and is generally our primary means to manipulate a browser's DOM. While ostensibly governed by W3C standards, we all know each browser contains variations and idiosyncrasies. Script code that works on one version of a browser might not work on a different version of the same browser. These scenarios cause a lot of pain in testing and re-writing of JavaScript code. This pain isn't the language's fault – we will never see the day when all browsers implement web standards with 100% accuracy. JavaScript is an accessible language. We don't need special tools or compilers. We can view the source code of any page on the Internet and copy the script code for our own purposes – and many people do. Of course, not everyone who uses JavaScript is a software developer with an eye for good code. All sorts of people use JavaScript, and all sorts of ugly JavaScript code perpetuates itself on the Internet. Even software developers (the author included) have taken a quick and dirty approach to writing JavaScript. It's only script code, after all, and just slapping the code into a text editor to get the desired result is all we need. It's not until we have to untangle a mess that we realize a more disciplined approach would have ultimately saved us time. With all of these problems in the JavaScript environment – why do we still want to torture ourselves by writing JavaScript code? Over the last few years, the amount of JavaScript code you'll find in the typical web application has surged. There are a couple good reasons for the surge: If you want to write an application that will reach as many users as possible, then you'll be writing a web application. You can reach users on Windows, Macintosh, Linux, and hundreds of other platforms on devices both large and small. How will you make a web application interactive? You'll use JavaScript. Your users won't have to install a runtime, or an ActiveX control, or download some interpreting engine. They'll install a web browser that includes JavaScript support, as so many do, and they'll happily use your application. JavaScript is the most ubiquitous programming language on the planet. As the demand for JavaScript code has increased, the frameworks and libraries of well-tested and robust JavaScript code have begun to emerge. Many of these frameworks abstract away the browser idiosyncrasies we discussed earlier, and can greatly reduce the amount of time we invest in writing and debugging cross-platform JavaScript code. Here are some of the most popular frameworks: We've also begun to see the emergence of proven practices and design patterns. The practices put the object oriented features of JavaScript to good use. What's that? You didn't know JavaScript was object oriented? We might not have applied OOP practices to JavaScript code over the last few years, but the capability does exist. JavaScript does have an object data type – but these objects can behave differently from the objects we create in C# and VB code. In C# and VB we create new objects by telling the runtime which class we want to instantiate. A class is a template for object creation. A class defines the properties and methods an object will have, and these properties and methods are forever fixed. We can't manipulate an object by adding or removing properties and methods at runtime. In JavaScript there are no classes, so we have no template for object creation. Then how do properties and methods become part of an object? One approach is to dynamically create new properties and methods on an object after construction. To create a new property, all we need to do is assign a new value for a property. The following code creates a new object, then adds an x and y property to the object. Finally, the script writes the values of the two new properties into an area in the HTML document. var p = new Object();p.x = 3;p.y = 5;message.innerHTML = p.x + "," + p.y; JavaScript objects are entirely different from C# and VB objects because they are ultimately a collection of name and value pairs. We can access an object's values using the dot operator "." followed by the name of the value. A value isn't constrained to holding simple integer types as we have shown in the first example, but could hold an array, function, or even other object. If you are thinking a JavaScript object sounds like a Dictionary from the .NET framework class library, then you are heading in the right conceptual direction. A JavaScript object is similar to a Dictionary<string, object> in the sense we can associate any arbitrary piece of data with any arbitrary string. In fact, there is an alternate syntax for accessing the values inside an object which makes the object look exactly like a dictionary. Instead of using the . operator to access a value, we can use the [] operator. The [] operator is a common sight when we work with collections like arrays, dictionaries, and hashtables. Let's rewrite our first example using the [] operator. var p = new Object();p["x"] = 3;p["y"] = 5;message.innerHTML = p["x"] + "," + p.y; The above piece of script produces the same result as our first script. It creates a new object, then adds x and y properties to the object, then writes out the values. Note that if we access a property that does not exist, we'll get back a value of "undefined". For instance, the line of code "alert(p.z);" would force a dialog box to appear with the string "undefined" inside. We can also add functions into the collection of values inside an object. Functions associated with an object become methods of the object. The following code sample shows how to create and use a method with the name of "print". var p = new Object();p["x"] = 3;p.y = 5;p["print"] = function() { message.innerHTML = p.x + "," + p.y; }p.print(); Notice we alternate use of the . operator and the [] operator. We can use these two operators interchangeably, for the most part, to create and access an object's properties and methods. Sometimes these operators lead to confusion, because it's not clear if a particular piece of code is trying to create new properties on an object, or if it's trying to set existing properties to new values. Fortunately, there is a third syntax available that makes our intent explicitly clear. The object literal syntax of JavaScript allows us to create an object and specify its properties using shorthand. The syntax uses a comma-separated list of name and value pairs, where the name and the value themselves are separated by a colon. Let's rewrite our code and create our object using this object initialization syntax. var p = { x : 5, y : 3, print : function() { message.innerHTML = p.x + ',' + p.y; }}p.print(); In the above code it becomes clear where object initialization begins and ends. Also note that we can nest object literals, and that the property values inside the object literals do not need to be constants – we can use any legal JavaScript expression. The following code will contains a nested object (address), and assigns the current date to a new createdDate property. var person = { name: "Scott Allen", createdDate: new Date() website: "OdeToCode.com", address: { state: "MD", postalCode: "21044" },};alert(person.address.state);alert(person.createdDate); The object literal syntax is popular because of its explicit intent and compact size. If you look at the source code for many of today's popular JavaScript frameworks, you'll see they are using object literals inside. Frameworks, however, aren't the ones using object literals. JavaScript Object Notation (JSON) is a lightweight data-interchange format based on a subset of the object literal syntax. Technically, JSON is a stricter version of the object literal syntax. For example, string literals must be enclosed in double quotes – no single quotes are allowed. JSON allows JavaScript to exchange data over the network (typically with the XmlHttpRequest object) and interoperate with other applications. Many web service providers offer JSON as a serialization format and as an alternative to XML. When our JavaScript contacts the web service, the web service will return its data in JSON. There is no need for our code to manipulate XML data with an XML API - instead our code can use JavaScript's eval statement to convert JSON into an object graph. var jsonString = "({ x : 3, y: 5 })";var p = eval(jsonString);alert(p.x + ',' + p.y); JSON is becoming hugely popular on the web. JSON is human readable and easily consumable in JavaScript. Also, exchanging data with JSON typically results in smaller payloads than using XML. ASP.NET AJAX includes a JavaScriptSerializer class to use JSON on the server-side in managed code. With our brief diversion into JSON complete, let's return to the topic of object oriented JavaScript. So far we've learned the following: This is useful information for constructing objects, but it's only a starting point. To get to the next level of abstraction, we'll need to add a second piece of knowledge: #2 is what we will discuss in the next topic. A JavaScript function is a chunk of executable code, but it's also a first class object. This is fundamentally different from methods in C# and Visual Basic. We can invoke methods in C# and VB, but we can't treat those methods as datatypes (although delegates and lamda expressions in C# make this area a little bit fuzzy). In JavaScript, we can manipulate functions using other JavaScript code, assign functions to variables, store functions inside arrays, nest functions inside other functions, and pass functions as a parameter to other functions. This might sound strange, so let's walk through a simple example. function add(point1, point2){ var result = { x : point1.x + point2.x, y : point1.y + point2.y } return result; } The above code defines a function named "add". The function expects two parameters, and expects that these two parameters will both have x and y properties that it can add together. It returns the result as a new object (created in object notation) with x and y properties. We could use invoke this function as in the following sample: var p1 = { x: 1, y: 1 };var p2 = { x: 1, y: 1 };// use our add functionvar p3 = add(p1, p2);alert(p3.x + "," + p3.y); The resulting dialog box will display "2,2". Technically, what we've done with the add function is create a new function object, and assigned the function object to a variable named add. We could take the same function object and assign it to different variables and invoke the function through those variables. function add(point1, point2){ var result = { x : point1.x + point2.x, y : point1.y + point2.y } return result; }var foo = addvar bar = addvar p1 = { x: 1, y:1 };var p2 = { x: 1, y:1 };// invoke add through foo variable// p3 should be 2,2var p3 = foo(p1, p2); // invoke add again through bar variable// 2,2 + 1,1 = 3,3p3 = foo(p3, p1); alert(p3.x + "," + p3.y); The resulting dialog box should now display "3,3". We can also assign a function object to an object property. As we noted before, this promotes the function to the status of "method". var point1 = { x: 3, y: 5, add: function(otherPoint) { this.x = this.x + otherPoint.x; this.y = this.y + otherPoint.y; } };var point2 = { x: 1, y: 1};// add 3,5 to 1,1point1.add(point2);// shows 4,6alert(point1.x + "," + point1.y); The first part of the code uses object notation to create an object with x, y, and add properties. The add property is a function object, and inside we've introduced the "this" keyword. Just as every instance method in C# has an implicit "this" parameter (and every instance method in VB has "Me" parameter), every JavaScript method has an implicit "this" parameter that represents the object through which the method was invoked. "this.x" will reference the x property of point1, because we invoke the add method using point1. What is a problem is that we have two "point" objects, but one has an add method and one does not. Remember, we are not defining classes like we would in C# or VB, we are simply creating objects and adding properties and methods on the fly. If we wanted to same add method in both point1 and point2, we could write the following code. function addPoints(otherPoint) { this.x = this.x + otherPoint.x; this.y = this.y + otherPoint.y;} var point1 = { x: 3, y: 5, add: addPoints};var point2 = { x: 1, y: 1, add: addPoints};// add 3,5 to 1,1point1.add(point2);// shows 4,6alert(point1.x + "," + point1.y); Now we've defined a function object and assigned the object to a variable named addPoints. We use addPoints to create new add methods in both the point1 and point2 objects. Does the "this" reference still work in addPoints? Yes it does, because "this" will still reference the object through which the method was invoked. "this" is a bit ephemeral in JavaScript, as we see later on, but we can can now invoke the add method on either the point1 or point2 object. This syntax is feeling uncomfortable, however. It looks as if we are trying to create a Point class that will define the properties and methods for all Point objects. But JavaScript doesn't have classes, so that would be a dream, right? We'll forever need to include all this object literal code every time we need a point object, right? Let's hope we never move into 3 dimensions where the definition of our point objects will change. Fortunately, there is a better solution. In JavaScript, a constructor function works in conjunction with the new operator to initialize objects. A constructor function can improve our previous code, because we can use the function to initialize every object we want to use as a Point. function Point(x,y){ this.x = x; this.y = y;}var p1 = new Point(3,5);var p2 = new Point(4,6);// shows 3,5alert(p1.x, p1.y); Constructor functions are just regular functions. It's just we've designed the function to be used with the new operator. By convention, we generally capitalize constructor functions to make other programmers aware of their significance. When we use the new operator with the Point function, the new operator will first create a new object. The new operator then invokes the Point function and passes a reference to the newly created object in the implicit "this" parameter. Inside the Point function we are creating new name/value pairs using the parameter values passed to the function. We can also create methods on an object inside a constructor function. function Point(x,y){ this.x = x; this.y = y; this.add = function(point2) { this.x += point2.x; this.y += point2.y; }}var p1 = new Point(3,5);var p2 = new Point(4,6);p1.add(p2);// shows 7,11alert(p1.x + ',' + p1.y); This approach works well, but there is an alternate approach we can use which is more in favor today. To understand this approach, we'll need to introduce a new piece of knowledge. Let's review the first two: Now for number 3: Let's talk about prototypes. Prototypes are a distinguishing feature of the JavaScript language. C#, Visual Basic, C++, and Java are all examples of class-based programming languages. To create objects, we must first write a class that defines fields, properties, methods, and events. When we create a new object, we are creating an instance of that class. In JavaScript, there are no classes. JavaScript is a prototype-based programming language. Every object has a prototype property that references its prototype object. Any properties and methods that are a part of an object's prototype will appear as properties and methods of the object itself. Remember that every function is an object, and every object references a prototype object. That means every constructor function references a prototype object. This is extremely useful when used in conjunction with the "new" operator, because of steps taken by the new operator: The above 3 steps mean that all objects created by a constructor function will have the same prototype – the prototype object for the constructor function. If we can modify the constructor function's prototype object, we will modify all objects the constructor function ever creates (or has already created). You can almost think of every object as inheriting from it's prototype, because it will include all the properties and methods defined by its prototype. I say "almost" because this thinking can be dangerous in some edge cases. Fortunately, there is an easy syntax we can use to add new properties and methods into a prototype object. Remember, all objects in JavaScript are dictionaries, and a prototype object is no exception. We can modify an object' s prototype simply by referencing it's prototype property, and we can add properties and methods to that prototype object. Let's rewrite our Point "class" one more time. function Point(x,y){ this.x = x; this.y = y;}Point.prototype.add = function(point2){ this.x += point2.x; this.y += point2.y;}var p1 = new Point(3,5);var p2 = new Point(4,6);p1.add(p2);// shows 7,11alert(p1.x + ',' + p1.y); The methods and properties we add to Point.prototype will be shared by all objects that are constructed from the Point constructor function. When we add methods to an object using the constructor function – each object gets a new property referencing a function object, so the prototype approach is more efficient (shared function objects) as well as being a little easier to read. This prototype approach is used by many of today's JavaScript frameworks. The topics we've discussed so far put us close to "simulating" classes in JavaScript. There are just a couple more topics we need to introduce before wrapping up. One topic is encapsulation. In JavaScript, every name/value pair we add to an object becomes a public property. There are no keywords in JavaScript to restrict accessibility (like the internal, protected, and private keywords in C#). Nevertheless, we can simulate private members. Douglas Crockford published an article "Private Members In JavaScript" that demonstrates how to add private members to a JavaScript object. Information hiding is an important technique in object oriented programming, and many JavaScript toolkits use Crockford's approach to private members. Private members have to be made in an object's constructor function. Both local vars and parameters are eligible to become private members using a closure. A closure in JavaScript is an inner function that references a local var or parameter in its outer function. Those local variables and parameters, which typically go out of scope when the outer function finishes execution are now "enclosed" by the inner function, which can continue to reference and use those variables. Let's re-write our sample once more, this time providing public "get" and "set" accesors for our points. function Point(x, y){ this.get_x = function() { return x; } this.set_x = function(value) { x = value; } this.get_y = function() { return y; } this.set_y = function(value) { y = value; }}Point.prototype.print = function(){ return this.get_x() + ',' + this.get_y();}var p = new Point(2,2);p.set_x(4);alert(p.print()); Client code can no longer access the x and y values of a point object directly. Instead, the code has to go through the set_ and get_ methods. Namespaces are crucial for avoiding type name collisions, which can be a bad thing in JavaScript. Unlike a compiled language like C# or VB, where a type name collision will result in a compiler error and an un-shippable product, in JavaScript you can still ship the code and might not find out about the collision until it's too late. JavaScript will happily overwrite one value with another. Since we are now including JavaScript code from all over the place, the practice of using namespace is important. There is just one problem. JavaScript doesn't support namespaces. This is ok, because we can "simulate" namespace using objects. Let's put our "Point class" into a Geometry namespace. var Geometry = {}Geometry.Point = function(x,y){ this.x = x; this.y = y;}Geometry.Point.prototype ={ print: function() { return this.x + ',' + this.y; }}var p1 = new Geometry.Point(5,2);alert(p1.print()); Essentially, we are adding our constructor function to the Geometry object. By adding other constructor functions (Rectangle, Square, etc), we could keep all of our types inside Geometry and not pollute the global namespace. Most JavaScript frameworks use a similar technique. This article presented three key pieces of knowledge: Those are three fundamental facts about JavaScript that also make JavaScript different from mainstream CLR languages like C# and VB.NET. Embracing and internalizing these differences will put you ahead of the game in understanding modern JavaScript frameworks, toolkits, and libraries. This article is by K. Scott Allen. Questions? Comments? Bring them to my blog.
http://odetocode.com/Articles/473.aspx
crawl-001
refinedweb
3,750
65.22
Regex.Replace Method (String, MatchEvaluator, Int32) In a specified input string, replaces a specified maximum number of strings that match a regular expression pattern with a string returned by a MatchEvaluator delegate.. - count - Type: System.Int32 The maximum number of times the replacement will occur., Int32) the first countMatch objects deliberately misspell half of the words in a list. It uses the regular expression \w*(ie|ei)\w* to match words that include the characters "ie" or "ei". It passes the first half of the matching words to the ReverseLetter method, which, in turn, uses the Replace(String, String, String, RegexOptions) method to reverse "i" and "e" in the matched string. The remaining words remain unchanged. using System; using System.Text.RegularExpressions; public class Example { public static void Main() { string input = "deceive relieve achieve belief fierce receive"; string pattern = @"\w*(ie|ei)\w*"; Regex rgx = new Regex(pattern, RegexOptions.IgnoreCase); Console.WriteLine("Original string: " + input); string result = rgx.Replace(input, new MatchEvaluator(Example.ReverseLetter), input.Split(' ').Length / 2); Console.WriteLine("Returned string: " + result); } static string ReverseLetter(Match match) { return Regex.Replace(match.Value, "([ie])([ie])", "$2$1", RegexOptions.IgnoreCase); } } // The example displays the following output: // Original string: deceive relieve achieve belief fierce receive // Returned string: decieve releive acheive belief fierce receive Available since 8 .NET Framework Available since 1.1 Portable Class Library Supported in: portable .NET platforms Silverlight Available since 2.0 Windows Phone Silverlight Available since 7.0 Windows Phone Available since 8.1 The regular expression \w*(ie|ei)\w* is defined as shown in the following table. The regular expression pattern ([ie])([ie]) in the ReverseLetter method matches the first "i" or "e" in the diphthong "ie" or "ei" and assigns the letter to the first capturing group. It matches the second "i" or "e" and assigns the letter to the second capturing group. The two characters are then reversed by calling the Replace(String, String, String) method with the replacement pattern $2$1.
https://msdn.microsoft.com/en-us/library/e47f3dkc(v=vs.110).aspx
CC-MAIN-2017-34
refinedweb
328
50.84
Matplotlib: interactive plotting¶ Interactive point identification¶ I find it often quite useful to be able to identify points within a plot simply by clicking. This recipe provides a fairly simple functor that can be connected to any plot. I've used it with both scatter and standard plots. Because often you'll have multiple views of a dataset spread across either multiple figures, or at least multiple axis, I've also provided a utility to link these plots together so clicking on a point in one plot will highlight and identify that data point on all other linked plots. import math import matplotlib.pyplot as plt class AnnoteFinder(object): """callback for matplotlib to display an annotation when points are clicked on. The point which is closest to the click and within xtol and ytol is identified. Register this function like this: scatter(xdata, ydata) af = AnnoteFinder(xdata, ydata, annotes) connect('button_press_event', af) """ def __init__(self, xdata, ydata, annotes, ax=None, xtol=None, ytol=None): self.data = list(zip(xdata, ydata, annotes)) if xtol is None: xtol = ((max(xdata) - min(xdata))/float(len(xdata)))/2 if ytol is None: ytol = ((max(ydata) - min(ydata))/float(len(ydata)))/2 self.xtol = xtol self.ytol = ytol if ax is None: self.ax = plt.gca() else: self.ax = ax self.drawnAnnotations = {} self.links = [] def distance(self, x1, x2, y1, y2): """ return the distance between two points """ return(math.sqrt((x1 - x2)**2 + (y1 - y2)**2)) def __call__(self, event): if event.inaxes: clickX = event.xdata clickY = event.ydata if (self.ax is None) or (self.ax is event.inaxes): annotes = [] # print(event.xdata, event.ydata) for x, y, a in self.data: # print(x, y, a) if ((clickX-self.xtol < x < clickX+self.xtol) and (clickY-self.ytol < y < clickY+self.ytol)): annotes.append( (self.distance(x, clickX, y, clickY), x, y, a)) if annotes: annotes.sort() distance, x, y, annote = annotes[0] self.drawAnnote(event.inaxes, x, y, annote) for l in self.links: l.drawSpecificAnnote(annote) def drawAnnote(self, ax, x, y, annote): """ Draw the annotation on the plot """ if (x, y) in self.drawnAnnotations: markers = self.drawnAnnotations[(x, y)] for m in markers: m.set_visible(not m.get_visible()) self.ax.figure.canvas.draw_idle() else: t = ax.text(x, y, " - %s" % (annote),) m = ax.scatter([x], [y], marker='d', c='r', zorder=100) self.drawnAnnotations[(x, y)] = (t, m) self.ax.figure.canvas.draw_idle() def drawSpecificAnnote(self, annote): annotesToDraw = [(x, y, a) for x, y, a in self.data if a == annote] for x, y, a in annotesToDraw: self.drawAnnote(self.ax, x, y, a) To use this functor you can simply do something like this: x = range(10) y = range(10) annotes = ['a', 'b', 'c', 'd', 'e', 'f', 'g', 'h', 'i', 'j'] fig, ax = plt.subplots() ax.scatter(x,y) af = AnnoteFinder(x,y, annotes, ax=ax) fig.canvas.mpl_connect('button_press_event', af) plt.show() This is fairly useful, but sometimes you'll have multiple views of a dataset and it is useful to click and identify a point in one plot and find it in another. The below code demonstrates this linkage and should work between multiple axis or figures. def linkAnnotationFinders(afs): for i in range(len(afs)): allButSelfAfs = afs[:i]+afs[i+1:] afs[i].links.extend(allButSelfAfs) subplot(121) scatter(x,y) af1 = AnnoteFinder(x,y, annotes) connect('button_press_event', af1) subplot(122) scatter(x,y) af2 = AnnoteFinder(x,y, annotes) connect('button_press_event', af2) linkAnnotationFinders([af1, af2]) I find this fairly useful. By subclassing and redefining drawAnnote this simple framework could be used to drive a more sophisticated user interface. Currently this implementation is a little slow when the number of datapoints becomes large. I'm particularly interested in suggestions people might have for making this faster and better. Handling click events while zoomed¶ Often, you don't want to respond to click events while zooming or panning (selected using the toolbar mode buttons). You can avoid responding to those events by checking the attribute of the toolbar instance. The first example below shows how to do this using the pylab interface.() If your application is in an wxPython window, then chances are you created a handle to the toolbar during setup, as shown in the {{{add_toolbar}}} method of the embedding_in_wx2.py example script, and can then access the {{{mode}}} attribute of that object (self.toolbar.mode in that case) in your click handling method. Section author: AndrewStraw, GaelVaroquaux, Unknown[99], AngusMcMorland, newacct, danielboone
http://scipy-cookbook.readthedocs.io/items/Matplotlib_Interactive_Plotting.html
CC-MAIN-2018-26
refinedweb
745
51.95
. If you want more background information on best practices and practical use cases, definitely read “The Mystery of CSS Sprites: Techniques, Tools and Resources.” If you’re the defensive type, I would recommend “CSS Sprites: Useful Technique, or Potential Nuisance?,” which discusses possible caveats. I won’t take a stance on the validity of CSS sprites. The aim of this article is to find out why people still find it difficult to use CSS sprites. Also, we’ll come up with a couple of substantial improvements to current techniques. So, start up Photoshop (or your CSS sprite tool of choice), put on your LESS and Sass hats, and brush up your CSS pseudo-element skills, because we’ll be mixing and matching our way to easier CSS sprite implementation. PresetsScripts directory, and choose File → Scripts → Sprite Grid from the Photoshop menu..) ↑ Back to topShare on Twitter MethemerApril 11, 2012 5:18 am Great article, but when talking about LESS, SASS and Spriting and not mentioning the beauty of automagical Compass Sprites seems a bit off. I mean at least for users of SASS – it’s no problem turning Compass on (whether you’re using an app to compile or the terminal), and when compass is on – it automagically generates a sprite from all your images in the specified directory. I never even open photoshop to do it anymore. Rizky SyazuliApril 11, 2012 8:51 am agree.. with Compass+Sass it only takes two lines of code. here’s the tutorial for creating sprites with Compass: stanApril 12, 2012 5:28 am Yes, not mentioning compass is a big miss here Niels MatthijsApril 12, 2012 9:39 am I did read up on SASS/Compass and while very interesting, it’s not a methodology, rather an implementation on a single framework. Definitely handy when using SASS (and it could’ve been mentioned in the article as an afterthought), but far too flimsy to build an entire workflow on. The thing is that the point of this article still stands even when you leave the whole LESS thing out of it. Maybe the focus on the LESS/SASS thing shifted the focus of the article in the wrong direction, then again people who prefer LESS could still benefit from the functions. Thanks for the heads-up though! YojimboApril 11, 2012 5:19 am I have to manage a big sprite, and I found a solution which perfectly fits my needs I have several photoshop files, some are messy, but it’s not a problem, then i slice everything i want. So I end with a lot of little images. Then, I import all the pack in this shiny little app, and TADA : it compacts me all the images to make the most little output image possible, and it create the CSS file linked to it (each icon with a classname equal to the slice name). (Note that the app is still a bit crashy, I hope it will be resolved, but my output files are ok each time) LeimiApril 11, 2012 5:29 am Nice post :) The mixins are indeed a nice addition in order to stop doing the maths. For Sass+Compass users though, working with sprites is really a piece of cake, and all those mixins are still too much work ;) With Compass, you basically just have to put all the images you want in the same folder and you’re good to go. Compass will dynamically build the sprite as you want and will know for you the position of each image in the sprite. A bunch of mixins and helpers are already at your disposal to work. This is one of the key features of Sass/Compass compared to Less for me. CarsonApril 11, 2012 5:33 am I strongly disagree. Making extra large images and relying on a CSS preprocessor in order to more easily remember a few numbers is overkill. The load time along is enough to kibosh this approach. All you should need is the info panel in Photoshop and a few guides to do the job. Don’t over-complicate things. AnahkiasenApril 11, 2012 7:41 am ^ That’s only true if you compile your less file in real time in production which I don’t think is a good idea. And as said below, spriting your image in Photoshop removes the flexibility of having your images as they are. —- Nice article but as mentioned, it’s kind of obvious that you’re a LESS developper before being SASS one because SASS and Compass offer something much more simpler than that. You put your images in a folder, @import that folder in your sass code, and it will automatically generate a sprite from that folder, with corresponding mixins to call each image by their name. Like this : @import interface/*.png button { +interface-sprite(‘button’) } And it will automatically select the right spot in the sprite to include. It even manages hover and active states automatically. You can configure your generated sprite as you wish etc. Really, LESS gained a lot of features recently but on some fields it will never be as powerful as what both SASS and Compass combined offer. Take a look at the Compass Sprite documentation to see more what I mean : PaulApril 12, 2012 12:46 am Compass sounds cool… I really don’t understand why they wouldn’t want to support LESS users as well :( AnahkiasenApril 12, 2012 1:41 am Because SASS has built-in conditions and loops, such as @if, @for, @while, @else, @each, etc, as well as lists and associative lists. Those are at the core of Compass and allow its flexibility, without them none of it would exist, and unfortunately LESS has none of those syntaxes. KumoApril 12, 2012 2:24 am I have to agree with you on this. Ok it requires a bit of concentration but it’s not such a hassle to tae the lelection tool and make a rectangle from the top left of your canevas to the top left of your icon. — I’ts an interesting technique indeed and it might be useful on some cases. But in my own opinion it gets thing more complicated that they really are. RonnieSanApril 12, 2012 8:31 am If you’re building your sprite in Photoshop, you can create slices around each image in your sprite. Then when you open the dialog box for each slice, there are boxes with the x-coordinate, y-coordinate, width, and height of the slice. In your CSS, just make the x and y coordinates negative. Done. You don’t have to measure each time or try to remember all the coordinates or sizes. MasumaSeptember 5, 2012 9:19 am It was actually worse than that, too. I made a thared to introduce myself and show my sprite, and immediately got comments saying PUT IT IN THE COMPLETED SPRITES FOR USE THREAD , which doesn’t make sense, and was done so without grace or so much as a hello a You get the picture. Oh well. Seems like the developers who actually make the game are really nice though, just the forum-goers who aren’t so friendly. Christian EngelApril 11, 2012 5:49 am We had almost the same experience as you described in your first paragraph. It ist one of the most unpleasant tasks to build a spritemap with photoshop and a editor tool, since you always have to switch between them and keep positions and dimensions in mind. Especially looking up positions and dimensions of the sprites is a hard job with photoshop. So we decided to leave photoshop completely out of the process. You should definitely have a look at Its a free HTML5 web-app and makes building CSS spritemaps finally a fun job to do. I really wonder why you had not heard about it in the last weeks! Give it a try and tell us what you think about it ;) Julien JollyApril 12, 2012 12:49 am Have a look at what? RonnieSanApril 12, 2012 8:32 am I think SM is removing links from comments or something. There are a few comments missing references to what they are talking about. Christian EngelApril 17, 2012 7:30 am Yeah, I originally linked to spritepad.wearekiss.com – the link got removed, I don’t know why… Alex BurrApril 11, 2012 5:52 am Great article. Unfortunately I have been receiving errors while trying to use the sprite grid script: Error 21: undefined is not an object (Line: 7) Is it intended for a particular version of Photoshop? Niels MatthijsApril 11, 2012 8:55 am We’re looking into it, but the function should be supported from CS2 and up. It could be that older versions of PS might be giving us trouble though. JasonApril 12, 2012 4:12 am I’m using CS4 Jon-Paul LeClairApril 12, 2012 7:20 am CS3 Here… looking at the CS3 scripting guide… there’s no mention of the “.guides” namespace. So the linked script won’t work. James BurkeOctober 10, 2012 3:15 pm I updated the Photoshop script to work in at least CS4 (I don’t have any older versions to try out, sorry). Disclaimer: None of the code is mine, I just hacked a new function together Neo MudalyOctober 12, 2012 9:43 am Didn’t have luck with Sprite Grid either, nor the @James Burke solution. Here’s a tidy extension for Photoshop grids: JasonApril 11, 2012 5:59 am I’m getting an error when running the photoshop script: Error 21: undefined is not an object. Line:7 -> doc.guids.add(Direction.VERTICAL, i * columnWidth); Peter MüllerApril 11, 2012 6:03 am Having your original images in a sprite sheet during development is an ugly hack that makes a lot of things more difficult for both the graphics designer and the web developer. The individual images are suddenly no longer available, making reuse impossible, making version control difficult, making image optimization more difficult etc. Further more you will always get in trouble with to much unused space in the sprite sheet, and having to do a lot of reorderings when some images are replaced, change dimensions etc. The real way to solve this problem is to keep using the original individual and non-space-optimized images during development and to introduce a build step that generates your sprite sheet and replaces the relevant css to use the generated sprite sheets. This gives you a lot of extra freedom because you no longer have to think about background offsets, sprite sheet rearrangements, optimization into seperate reduced color palette sprites etc. All images will be available in their original unaltered an unoptimized form, which means you can experiment with optimization in your build step without changing anything in your actual image or css. And of course it’s much easier to debug. Since most developers that need to do hard core optimizations, like sprites to reduce http requests, will have a build step for crushing their css and javascript anyway, adding automation of spriting into the mix is not that big of a leap. But it makes a huge difference to the work environment for the designer and the developer during implementation. The company I work for has open sourced the build system we developed for creating large web applications. Apart from all the optimizations that you would expect with minification and bundling, it also adds automated spriting, image optimization with pngcrush, pngquant, optipng and allows you to define several sprite sheets so you could optimize for reduced color palettes with pngquant, thus taking total image size much lower than a single sprite sheet ever would. The recipe for these optimizations is defined through vendor specific css properties, so it’s actually valid css, and thus you can debug on the unbuilt files during development. As an added bonus the build system will also automatically base64 encode and inline css images in case they are only referenced once and are within a sane boundary in dimensions, thus sometime even removing your need for a sprite sheet completely. If anyone is interested, take a look at The post process optimization syntax hasn’t been documented yet, but the spriting syntax is described in the readme. The complete build system the spriting system is a part of is also available, and can be highly recommended if you haven’t already implemented an automated optimization build step into your workflow An example of a page that uses all aspects of this automated spriting and image optimization can be found on the default page my employer puts on new blank web spaces. This one is blank for now Rizky SyazuliApril 11, 2012 8:57 am seems that the link to the open source repo you’ve mentioned has been stripped from the comment. your company’s build system sounds awesome.. Peter MüllerApril 11, 2012 11:41 pm Indeed it seems so. I wonder how they expect comments to be useful if noone can link to external resources. My companys build system is called Assetgraph-builder, and is the first hit for me on google and on github. The sprite part of it is called Assetgraph-sprite. That should hopefully be enough for you guys to find the resources with a little manual work. The example page that I tried linking to is jimmyborg dot net Rizky SyazuliApril 12, 2012 2:22 am thx Peter, i found it. lots of interesting stuff there. will try it out soon.. Shane MurphyApril 11, 2012 6:06 am Nice little tutorial… liking the use of LESS for the calculations :) I found recently, which is a free tool that creates sprites and css automatically which I thought was pretty cool. Aljan ScholtensApril 11, 2012 6:08 am Hi, nice post, I’ve always worked with image sprites, but lately I’m having serious troubles with sprites because I work entirely responsive, working with em units and make the images beautiful on retina displays. What would you suggest to do for me and a lot of other designers? :) WilliamApril 11, 2012 6:31 am I think you could use the background-size property but it isn’t supported on some browsers. Aljan ScholtensApril 19, 2012 11:03 pm If I would use background-size on a sprite it would just break, it would show all images in the sprite if I’m using cover or 100%. I also think every screen will get more pixels in the near future (like Retina displays), so this should set us thinking about sprites, larger images (standard?) etc. Jaaon LydonApril 11, 2012 8:53 am Aljan nailed it for me. I used to use Sprites for everything, but with responsive and mobile sites, I find them to make the process much harder. background-size is money, but super complicated with sprites. Aljan ScholtensApril 19, 2012 11:04 pm True, the best way is to avoid images ;-) Frank BroersenApril 11, 2012 6:30 am Why not let SASS Generate the complete sprite for you? Felix EggbertApril 11, 2012 6:38 am Why not use a tool like that makes things easy…? Mike RobinsonApril 11, 2012 6:47 am Nice combination of LESS and advanced planning, it solves a problem a lot of designers seem to struggle with. Have you also considered reviewing automated CSS Sprite generators? While they tend to be less efficient, they do allow you to use traditional layout methods and still gain the majority of the performance boost on production. Kumsal ObuzApril 11, 2012 8:33 am Fear not and use this It generates CSS file for you out of your images. That’s right, you don’t have to right your own CSS. Even if you change your pictures’ size, remove/add them to the collection, you can easily update your previously exported CSS. Niels MatthijsApril 11, 2012 9:03 am As a general reaction to automatic sprite generators: I’ve tried a few, but found none that could output decent css rules. I’m not really sure about the SASS implementation, but if it could dynamically find and rewrite the background rules that would be impacted by the sprite while at the same time generating the sprite from a specific folder, then yeah … that sounds pretty cool. Now it seems to rely on functions and file names, which is promising, but still somewhat of a drag. Peter Müller: sounds like a great solution, I will definitely look into it. One thing to check is whether the setup of the system is actually worth the trouble and whether it integrates will in existing workflows. hdennisonApril 11, 2012 9:29 am I’ve been using Atlas CSS for this type of things, and it works like a charm. Just put all your icons on a folder, run the Photoshop script, and it generates a CSS with the coordenates and size of each icon, using the original filename as CSS selector. Have to add some icons? No problem, add them to the folder, run the script again, and copypaste the generated CSS, no more dealing with coordenates or sizes. If someone ports this to Fireworks, I’ll be the most happy developer in the world :P P.S. If you don’t want to have all the icons on separate files, you can use Layer Cake to generate them from your unique PSD file: P.S. 2 Once again, If someone ports Layer Cake, or even combine those two steps on one Fireworks script, I’ll be happy forever :P Daniel AdrianApril 11, 2012 10:38 am You could use SmartSprites Jacob RaskApril 11, 2012 1:08 pm If IE7 is not a concern, inline your images in your CSS as base64 encoded data URI:s instead of using sprites. There are multiple tools available, but I just keep a folder with image files, and a small bash script that creates an images.less/scss file with variable declarations such as $img-myicon: data:image/png;base64,.... I then use it like this: background-image: url($img-my-icon). Easy-peasy and no extra HTTP requests at all. Just remember to gzip your CSS. Gerd WippichApril 11, 2012 1:23 pm Someone deleted some of the links from the posts above (#3, #7, #11) – why? There are a lot of little helpers on the web I still don’t know, and now I never will… :-( VolkwardApril 12, 2012 2:50 am There’s also. It takes all your single graphics files and builds the sprite as well as the respective CSS-markup. Also SASS if you like. Volkward adumpaulApril 12, 2012 3:27 am Its really nice collection.Thank you. FrankApril 12, 2012 3:54 am Excellent idea to target specific squares within the sprite! I’ll definitely implement it. However, you lost me at LESS & SASS. I started reading and thought I would understand the core of the article, half way through i realized it was all about LESS & SASS. It’s tagged as “CSS” and “Essentials”. I disagree with the second one. It should at least be in title of the article. clyfishApril 12, 2012 4:19 am I created a CSS sprites generator aimed at area minimization using VLSI floorplanning algorithm. Have a try. BrunoFebruary 19, 2013 5:01 am Nice job! It’s very handful ! Is the link down? LeeApril 12, 2012 4:20 am I’m always using sprites but I’ve never really found it to be a problem except when I’ll occasionally reposition things within the sprite to make things a bit tidier. FabioApril 12, 2012 4:35 am Great Article NamelessApril 12, 2012 4:58 am With CSS3 becoming more usable in the web design world, is there still a need to use SASS and LESS? I played around with SASS and LESS a few times in some temp projects and loved them, but is it wise to still use them with CSS3 on the go? What about using SASS and/or LESS on larger websites (100 – 150 pages)? John FauldsApril 14, 2012 12:09 am Writing CSS3 rules is one of the main reasons you would use LESS or SASS because using mixins it can automatically create all the extra lines with vendor prefixes for you. Mark W. JarrellApril 12, 2012 5:45 am I like using SpriteRight. It’s a really easy mac app that builds sprite images and CSS for you. ThomasApril 12, 2012 6:46 am I use my own method: HTML: <span class="ai-gfx ai-gfx-45"> <img style="left:-45px; top:0" src="images/gfx.png" alt="Über Facebook kontaktieren"/> </span> CSS: span.ai-gfx { overflow: hidden; display: inline-block; position: relative; } span.ai-gfx.ai-gfx-45 { width: 45px; height: 45px; } span.ai-gfx img { position: absolute; } If you know CSS, you know what I mean. I wrap an inline-block element around a semantic correct img-tag. Then I position the img itself with inline css. - no semantic issues - image not as background, which is important for accessibility - option for background texture in the outline element - cleaner code (no selectors for each sprite) There’s an entry on my blog (still work in progress ;)): peterApril 12, 2012 8:23 am Your solution is to use inline-css? ThomasApril 12, 2012 8:59 am yes, for the coordinates of the sprite. And I have a good reason for that. There’s a direct connection between the img-tag and the result. I don’t have to search in the css file for coordinates. And I don’t produce such amount of code in my css file. Of course, you could make selectors for every image. Even in my solution. But I don’t think that’s the best way. There are other variations. You can make a matrix with your buttons on the y-axis and with the states them on the x-axis. Now, just add an img:hover rule with “left:XXpx” in your css et voila, you have hover states in the same image file. KaeligApril 12, 2012 1:14 pm Disable CSS. Look at the page. Now look at me. Look at the page again. Look at me. I’m on a horse (and you have images that break the content flow). Frank Villa-AbrilleApril 12, 2012 6:53 am I can’t believe you missed SmartSprites: CSS Sprite Generation Done Right,. This eliminates the need to manually update your sprites via Photoshop and CSS. SmartSprites allows you to generate your and create your sprites automatically – meaning you can continue to create and maintain your images individually. .” “In other words, no tedious copying & pasting to your CSS when adding, removing or changing sprited images. Just work with your CSS and original images as usual and have SmartSprites automatically transform them to the sprite-powered version when necessary.” peterApril 12, 2012 8:21 am I believe CSS sprites have not caught on due to overwhelming % of web developers DO NOT create their own images. Andy DaviesApril 21, 2012 11:17 am I reckon sprites haven’t caught on because many people don’t understand that the number of requests on a web page is a big problem… EyalApril 13, 2012 1:26 am Easily Generate your sprite from Adobe Fireworks: Set up graphics in place (If you are using ‘print software’ such as Photoshop or illustrator, you can import your lay-outed and arranged sprite image as one file), set up slices in fireworks according to your desired positions. Save file. Download: and extend fireworks. Open file, run installed command (commands->export) ‘Slices as CSS Sprite’ Open exported css file or copy and paste to your generated css Export image file in place Win! Need changes in sprite? well if you used Fireworks for creating the original images – just add /change / rearrange the graphics and hit the export command again… Another win ;) Nick BondApril 13, 2012 2:22 am Hi all, my big problem is that you want developers/designers to be developing and designing not optimizing. Let’s face it CSS sprites and inlining images is just the tip of the optimizing iceburg, and as we have seen these are time-consuming enough! My company has a product (Stingray Aptimizer) that auto-magically optimizes content using these techniques (and a multitude of others), so you don’t have to. Plus it keeps on optimizing even when the content changes etc. sign up for a free eval and take a look riverbed.com/us/products/stingray/stingray_aptimizer.php At present this is an IIS/SharePoint focused product (Microsoft is a reference here!), the next release will allow us to optimize all web content irrespective of the web server it is being served from. Nick Jorge BastidaApril 13, 2012 4:17 am Hey guys, Glue is a simple command line tool to generate CSS sprites, probably somebody will find this quite useful :D Enjoy! Flávio AraújoApril 13, 2012 10:45 am My “secret” to work with css sprites is Fireworks. I use to make a copy of my image file and import/open on fireworks. Then I use to create slices on fireworks (what I think it´s much better then photoshop to do this task). So, fireworks at the botton of the workspace have nice X Y Width Heigth tool. So clicking on each slice, i have the perfect XY/ X Y Width Heigth values to use om my CSS file. I Think this is (as far as I know) the best way to do this boring task. Michael ConnorsApril 17, 2012 7:36 am Agreed—Fireworks is the way to go. It solves most dimensioning/slicing issues. Anyone struggling with these should look to Fireworks! Mark SimchockApril 16, 2012 6:28 am For anyone who might be interested I threw together a proof of concept for responsive sprites: Naturally, I welcome any input, etc. There a couple limitations vs traditional sprites. But as mentioned it’s a proof of concept and the idea is still being developed. pomehApril 17, 2012 5:17 am You may also like this technique using SVG in place of CSS sprites Jean-DavidApril 21, 2012 4:01 am Use This site write for you the code to place the images of your sprite. Vignesh Nandha KumarApril 23, 2012 5:21 am I use a tool called *Spritemapper* and I wrote a wrapper script around to automate the process. So everytime I want to create a sprite, I simply put all the individual images in a separate folder and specify the folder name to the script. The script takes care of creating the sprite, even compressing it using trimage (a combination of optipng, pngcrush, advpng and jpegoptim) and generating the correspnding CSS. Once the script finishes, I just copy paste the generated CSS to my actual CSS file. Cool huh? You can get the script at Happy designing! KevinApril 27, 2012 12:30 pm I usually create my sites using individual images in dev mode, then when i’m just about ready to deploy to the server to go live I take all my individual images and drop them into spritepad from wearkiss () and make my CSS adjustments to suit. It may take about another half hour to make all the adjustments, but it sure beats the Photoshop > HTML editor round tripping. LinusApril 28, 2012 6:46 pm The new upcoming Fireworks CS6 will have the ability to export sprites images and CSS codes in various layouts. You can do this by using slices on objects and states as well. It saved you a lot of time to prepare the images. It will auto rearrange the images to save the extra space in between. The ability to work on a design and do CSS sprites any time is just amazing. This will change people to do CSS sprites. Watch Fireworks CS6 features tv.adobe.com/m/#!/watch/cs6-creative-cloud-feature-tour-for-web/introduction-to-fireworks-cs6/ Gustaff WeldonMay 6, 2012 4:30 am Just wondering, why noone mentioned that sprites in :after and :before will leave the <IE7 without icons but with text-indent applied. This should be mentioned in the samples as people often tend to just copy and paste code examples without consideration. What happened to progressive enhancement, I can't recall hearing IE7 is no longer relevant… David NemesMay 10, 2012 2:14 am Using sprites is easy and fun, just use . It’s a sprite generator. Norm OlsenMay 21, 2012 8:53 am I simply use SpriteRight, a Mac desktop store app that enables you to simply import an image, set any desired padding(if applicable), and when you export, it generates code that you can simply cut and paste into your web code (the only thing I found I had to do is adjust the path with the background URL to ensure it points to the proper directory where the sprite sheet is located on your server. No fussing around with coordinates and calculations. once I tried out SpriteRight, I don’t see the need for anything else IMO (of course, this is all subjective and a matter of personal taste and work flow preferences). Cheers PHPSessJuly 2, 2012 4:47 am There is a tool to find CSS sprite x and y coordinates SendilMarch 5, 2013 9:32 am For some reason I feel getsprityxy.com easy to use than sprite cow and it also works in IE (not a big deal though :) ) Jorge VargasJuly 10, 2012 11:15 am The best css position automatic spritecow.com Lucian AdamsonJuly 23, 2012 6:59 am For the users getting the “undefined” error relating to the guides.add portion of the Adobe Photoshop grid script: The script in question will only work for Adobe Photoshop CS5+ Bah.. so unfortunately those who are under CS5 are out of luck. Including me. :( SaraSeptember 18, 2012 10:25 am Hey! If your site is using php, then the math for the icons can easily be calculated and the coordinates echoed into resp. inline css… In a grid that’s super easy. In “wild” sprite it’s also easily done. The only thing you need is the sprite image, and if you change it just change the php code. In this manner you can easily manage several sprites. That is, let php pick up the sprite you’ll use for the moment… Sara, a smart eady going lady of Sweden Felix KleeJanuary 17, 2013 2:32 am The links to RSS, Facebook, Twitter, Newsletter in the top right corner of this page are empty when images are disabled. Concerning accessibility: Wouldn’t it be better to use regular images for these links? marioNovember 15, 2013 8:44 am just a tip: nonchiedercilaparolaNovember 21, 2013 11:28 am I feel stupid to ask this question but I have to do it: how is it possible to add text-number into working area? .” GottZFebruary 18, 2014 6:05 am sad that nobody mentions photoshop’s scripting abilities. just search for: css sprite javascript photoshop and you will find lots of export scripts that do the job.
http://www.smashingmagazine.com/2012/04/11/css-sprites-revisited-2/
CC-MAIN-2014-15
refinedweb
5,208
68.2
25 July 2011 07:51 [Source: ICIS news] By Fanny Zhang ?xml:namespace> GUANGZHOU (ICIS)--China The government has already imposed the price-based resource tax in 12 provinces, but now has plans to extend it to the entire country, they said. A draft of the proposed expansion was given clearance last week by Chinese premier Wen Jiabao for submission to the State Council, the country’s top decision-making authority. The extended resource tax is pending the council’s approval and may get passed soon, they added. The price-based resource tax on crude oil and natural gas was first introduced in Xinjiang, which has been levying the tax at a starting rate of 5% since 1 June 2010. The remaining 11 provinces that include Elsewhere in the country, the resource tax on crude oil and natural gas is currently calculated and paid on volume. Energy firms pay yuan (CNY) 14-30/tonne ($2-5/tonne) resource tax on crude and CNY7-15/cbm (cubic metre) on natural gas depending on the volume of output. Industry sources said that a nationwide resource tax reform would “absolutely” increase the costs of crude and natural gas in PetroChina will suffer the biggest rise in its costs as it is the largest oil and gas producer in the country, while Sinopec may sustain a far lesser impact as most of its crude comes from imports, a source from Sinopec said. “Independent refineries are going to benefit from the [tax] reform because they’re taking imported fuel oil as feedstock instead of domestic crude oil,” said the source. A PetroChina source said that the reform may increase the amount the company pays for resource tax by six times to over CNY30bn a year. “Petrochemical producers will raise their prices in view of more expensive raw material and consumers eventually will have to pay for the rise in costs,” said a source from China Petroleum and Chemical Industry Federation (CPCIF). “Choosing a time for [proceeding with the nationwide tax] is critical. [However,] we believe it won’t [be realised] for the next three months at least because of inflation,” said a CPCIF source. The country’s consumer price index, a gauge of inflation, hit a three-year high of 6.4% in June this year. The country may adjust other taxes, such as special proceeds on crude oil, to offset the hefty increase of the resource tax in order to ensure healthy development of domestic petrochemical and energy industries, he added. The PetroChina source said that the rates of special proceeds for crude oil are too high currently and should be reduced or removed when the resource tax reform is introduced. (
http://www.icis.com/Articles/2011/07/25/9479563/china-plans-to-extend-5-10-resource-tax-on-oil.html
CC-MAIN-2015-14
refinedweb
447
54.36
# Bad news, everyone! New hijack attack in the wild On March 13, a [proposal](https://www.ripe.net/ripe/mail/archives/anti-abuse-wg/2019-March/004585.html) for the RIPE anti-abuse working group was submitted, stating that a BGP hijacking event should be treated as a policy violation. In case of acceptance, if you are an ISP attacked with the hijack, you could submit a special request where you might expose such an autonomous system. If there is enough confirming evidence for an expert group, then such a LIR would be considered an adverse party and further punished. There were some [arguments against this proposal](https://www.ripe.net/ripe/mail/archives/anti-abuse-wg/2019-March/004601.html). With this article, we want to show an example of the attack where not only the true attacker was under the question, but the whole list of affected prefixes. Moreover, it again raises concerns about the possible motives for the future attack of this type. For the last couple of years, when our colleagues or we were describing BGP hijacking events, mostly MOAS conflicts were covered. MOAS (Multiple Origin Autonomous System) is a case when two different autonomous systems announce the conflicted prefixes with their respective ASNs in the origin of the AS\_PATH (first ASN in AS\_PATH, further in the text — origin ASN). However, [we can name 3 additional](https://pc.nanog.org/static/published/meetings/NANOG75/1892/20190219_Gavrichenkov_Four_Years_Of_v1.pdf) hijack types allowing a malicious party to manipulate AS\_PATH for different reasons, including bypassing today’s filtering and monitoring services. The famous [Pilosov-Kapela](https://we.riseup.net/assets/43591/defcon-16-pilosov-kapela.pdf) type of attack is last, but not least. A possible example of this attack is what we think we saw for the last few weeks, observing for the first time in the wild. Such an event could have an understandable nature and severe consequences. For those looking for the TL;DR version, scroll down to “The Ideal Attack.” ### Networking Background (for you to better understand the processes behind the incident) If you want to send a packet and have several prefixes in your forwarding table that includes a destination IP-address, you’ll use a route to prefix with the maximum length. If you have several different routes in your routing table for one prefix, you’ll choose the best one (under best path selection mechanism) to include in forwarding. Existing filtering and monitoring services try to analyze routes and make their decisions by looking on routes’ AS\_PATH attribute. BGP speaker can change this attribute to any value during the announcement. Simple adding of the owner ASN at the beginning of an AS\_PATH (as origin ASN) might be enough to bypass modern origin validation mechanism. Moreover, if there is any route from a target ASN to you, you can retrieve and use AS\_PATH from this route in your other announcements. Any validation check of **solely AS\_PATH** for this route would fail. There are several limitations that we want to mention. First — in case of prefix filtering by your upstream, your route can still be filtered out (even with the valid AS\_PATH) if the prefix does not belong to upstream’s customer cone. Second — valid AS\_PATH can become invalid if the crafted route is announced in wrong directions and thus start breaking routing policies. The last one — any route that has a prefix that is violating the ROA maxLength could be considered Invalid. ### The Incident A few weeks ago our user wrote us a complaint. We have seen routes with his origin ASN and /25 prefixes in it while the user claimed that he didn’t announce them. `TABLE_DUMP2|1554076803|B|xxx|265466|78.163.7.0/25|265466 262761 263444 22356 3491 2914 9121|INCOMPLETE|xxx|0|0||NAG|| TABLE_DUMP2|1554076803|B|xxx|265466|78.163.7.128/25|265466 262761 263444 22356 3491 2914 9121|INCOMPLETE|xxx|0|0||NAG|| TABLE_DUMP2|1554076803|B|xxx|265466|78.163.18.0/25|265466 262761 263444 6762 2914 9121|INCOMPLETE|xxx|0|0||NAG|| TABLE_DUMP2|1554076803|B|xxx|265466|78.163.18.128/25|265466 262761 263444 6762 2914 9121|INCOMPLETE|xxx|0|0||NAG|| TABLE_DUMP2|1554076803|B|xxx|265466|78.163.226.0/25|265466 262761 263444 22356 3491 2914 9121|INCOMPLETE|xxx|0|0||NAG|| TABLE_DUMP2|1554076803|B|xxx|265466|78.163.226.128/25|265466 262761 263444 22356 3491 2914 9121|INCOMPLETE|xxx|0|0||NAG|| TABLE_DUMP2|1554076803|B|xxx|265466|78.164.7.0/25|265466 262761 263444 6762 2914 9121|INCOMPLETE|xxx|0|0||NAG|| TABLE_DUMP2|1554076803|B|xxx|265466|78.164.7.128/25|265466 262761 263444 6762 2914 9121|INCOMPLETE|xxx|0|0||NAG||` *Announces example by the beginning of April 2019* NTT in a way for a /25 prefix route makes it especially suspicious. During the incident, NTT's LG knew nothing about this route. So yeah, some operator creates the whole AS\_PATH for these prefixes! Other speakers highlight one special ASN: AS263444. After looking at other routes with this ASN, we run into the following situation. `TABLE_DUMP2|1554076800|B|xxx|265466|1.6.36.0/23|265466 262761 263444 52320 9583|IGP|xxx|0|0||NAG|| TABLE_DUMP2|1554076800|B|xxx|265466|1.6.38.0/23|265466 262761 263444 52320 9583|IGP|xxx|0|0||NAG|| TABLE_DUMP2|1554076800|B|xxx|265466|1.23.143.0/25|265466 262761 263444 22356 6762 9498 9730 45528|IGP|xxx|0|0||NAG|| TABLE_DUMP2|1554076800|B|xxx|265466|1.23.143.128/25|265466 262761 263444 22356 6762 9498 9730 45528|IGP|xxx|0|0||NAG|| TABLE_DUMP2|1554076800|B|xxx|265466|1.24.0.0/17|265466 262761 263444 6762 4837|IGP|xxx|0|0||NAG|| TABLE_DUMP2|1554076800|B|xxx|265466|1.24.128.0/17|265466 262761 263444 6762 4837|IGP|xxx|0|0||NAG|| TABLE_DUMP2|1554076800|B|xxx|265466|1.26.0.0/17|265466 262761 263444 6762 4837|IGP|xxx|0|0||NAG|| TABLE_DUMP2|1554076800|B|xxx|265466|1.26.128.0/17|265466 262761 263444 6762 4837|IGP|xxx|0|0||NAG|| TABLE_DUMP2|1554076800|B|xxx|265466|1.64.96.0/20|265466 262761 263444 6762 3491 4760|IGP|xxx|0|0||NAG|| TABLE_DUMP2|1554076800|B|xxx|265466|1.64.112.0/20|265466 262761 263444 6762 3491 4760|IGP|xxx|0|0||NAG||` *You can try to guess what is wrong* It seems that someone took a prefix from the route, split prefix onto two parts and announce the route with same AS\_PATHs for these two prefixes. `TABLE_DUMP2|1554076800|B|xxx|263444|1.6.36.0/23|263444 52320 9583|IGP|xxx|0|0|32:12595 52320:21311 65444:20000|NAG|| TABLE_DUMP2|1554076800|B|xxx|263444|1.6.38.0/23|263444 52320 9583|IGP|xxx|0|0|32:12595 52320:21311 65444:20000|NAG|| TABLE_DUMP2|1554076800|B|xxx|61775|1.6.36.0/23|61775 262761 263444 52320 9583|IGP|xxx|0|0|32:12595 52320:21311 65444:20000|NAG|| TABLE_DUMP2|1554076800|B|xxx|61775|1.6.38.0/23|61775 262761 263444 52320 9583|IGP|xxx|0|0|32:12595 52320:21311 65444:20000|NAG|| TABLE_DUMP2|1554076800|B|xxx|265466|1.6.36.0/23|265466 262761 263444 52320 9583|IGP|xxx|0|0||NAG|| TABLE_DUMP2|1554076800|B|xxx|265466|1.6.38.0/23|265466 262761 263444 52320 9583|IGP|xxx|0|0||NAG|| TABLE_DUMP2|1554076800|B|xxx|28172|1.6.36.0/23|28172 52531 263444 52320 9583|IGP|xxx|0|0||NAG|| TABLE_DUMP2|1554076800|B|xxx|28172|1.6.38.0/23|28172 52531 263444 52320 9583|IGP|xxx|0|0||NAG||` *Example of routes for one of the split pairs of prefixes* Several questions arise. Who probed such hijack type? Did anyone accept these routes? Which prefixes were affected? So here starts our path of failure and yet another round of disappointment in the current state of the Internet health situation. ### Path of Failure However, first things first. How can we find which speakers accepted such hijacked routes and whose traffic is redirected already? We were thinking about /25 prefixes because “they just can not appear in the wild.” You can guess it — we were very wrong about it. It’s a very noisy metric, and routes with such prefixes can even appear from Tier-1 operators. Even NTT have around 50 prefixes that it propagates to its customers. On the other hand, this metric is terrible because these prefixes could be filtered out if the operator applies one of the BGP best practices, namely [small prefixes filtering](http://bgpfilterguide.nlnog.net/guides/small_prefixes/), in all directions. So we cannot use this method to find all of the operators whose egress traffic was redirected by an incident. Another wild thought was to look at the [ROV](https://datatracker.ietf.org/doc/rfc6811/). Specifically at routes with violation of the maxLength rule of a corresponding ROA. So we can look for some different origin ASN with Invalid status that was seen by the AS. Still, there is a “little” problem. The average (median and mean) of this number is around 150, and even if we filter out small prefixes, it will remain higher than 70. There is a simple explanation for this result: there are only a few operators who already apply a ROA based filter with “drop Invalid routes” policy on their ingress points, so wherever route with such violation appeared in the wild, it can spread in all directions. Last two approaches highlight accepted parts of the recent incident (because it was massive enough), but they are not generally applicable. However, maybe we can find an attacker? What are the general features of such an AS\_PATH manipulation? There are several assumptions: * The prefix did not appear previously in the wild; * Origin ASN (reminder: the first ASN in the AS\_PATH) is the valid one; * The last ASN in AS\_PATH is the ASN of the attacker (in case if his neighbor make a neighbor ASN check in all incoming routes); * The attack comes from one ISP. If all assumptions are correct, then in all malformed routes the hijacker ASN would be presented (besides origin ASN) and thus may be claimed as critical one. We can take any of previous feature (“/25” or “Invalid routes”), and for any ASN count, the number of different origin ASN whose routes matching this feature have a critical point in it. Among the true hijacker — AS263444, some others were highlighted even when we dropped routes with him from consideration. Why? Critical point can be critical even for reasonable routes. It could be a result of connectivity lack in some region or some limitations in our visibility. There is a way to detect an attacker, but only meeting several conditions, and only when the hijack is enormous to bypass monitoring thresholds. However, can we found out prefixes that suffered hijack regardless of all the above factors? Actually — yes. When hijacker creates a route with more specific, such prefix is not announced by the original ISP so, if you can somehow get a trustful list of all announced prefixes made by this operator itself, then for such operators you can make a comparison and find malformed routes with more specifics. We gather this list of prefixes from our BGP sessions. It’s simple enough — you're announcing not only the full view of routes that you see right now but also the list of all the prefixes that you are willing to announce in the world. Sadly enough, right now there are several dozens of our users, that don’t do the last part correctly. We’ll notify them in the nearest future and try to solve this problem. For all the others — you can join our monitoring system today. That is how we gather this particular data, represented in this article. If we return to the initial incident, the attacker and accepted parts were found out with critical points method. It’s strange, but not all the customers of AS263444 have seen these routes. Also, there is another strange point. `BGP4MP|1554905421|A|xxx|263444|178.248.236.0/24|263444 6762 197068|IGP|xxx|0|0|13106:12832 22356:6453 65444:20000|NAG|| BGP4MP|1554905421|A|xxx|263444|178.248.237.0/24|263444 6762 197068|IGP|xxx|0|0|13106:12832 22356:6453 65444:20000|NAG||` *A recent example of our address space hijack* When this hijacker creates more specific for our prefixes, it used crafted AS\_PATH. However, this AS\_PATH was not taken from any of our previous routes. We even don’t have a link with AS6762. We look at other routes in the incident: some of them were having real AS\_PATH that was used previously, but others — not, even if they look like real. There is no practical meaning in this change because in any case, traffic would be redirected to an attacker, but routes with bad AS\_PATH can be filtered out by ASPA or another verification mechanism. It raises questions about hijacker motives. As for now, we can’t claim that this incident was a planned attack. It could be, though. Let’s try to imagine a possible IRL situation. ### The Ideal Attack What do we end up with? You are a transit provider, announcing routes to customers. If they are multihomed, you will get only a part of their traffic — however, the more traffic — the more revenue. So, if you announce subnet prefixes of these routes with the same AS\_PATH, you’ll get the rest of the traffic. Also, the rest of the money. Could ROA help? Possibly yes, if you decide not to use [maxLength](https://datatracker.ietf.org/doc/draft-ietf-sidrops-rpkimaxlen/) at all. Also, you don’t have any ROA records with intersected prefixes in them. For some of the operators, it’s not a possible option. Looking at other security mechanisms, ASPA wouldn't help (because AS\_PATH is from the valid route) with this particular hijack. BGPSec is still a bad option due to deployment rate and remaining possibility of downgrading attack. So we have a clear profit for an attacker and the lack of security. A great mix! ### What should be done? The obvious and the most radical possible step — you can make a review of your current routing policy. Also, split your address space onto the smallest chunks (without intersections), that you have a desire to announce. Sign a ROA only for them without using a maxLength option. Then current ROV can and would save you from this attack. But again, for some operators, this is not a reasonable approach because of an exclusive use for more specific routes (all the problems in the current state of ROA and route objects would be described in one of our future articles). Also, we can try to monitor this kind of situation. To do so, we need ground truth about your prefixes. So you can contact us, set a BGP session with our collector and pass us information about your view of the Internet. We’ll highly appreciate if you send us a full-view (so we can find propagation scope of other incidents), but the whole list of routes with your prefixes should be enough for the beginning. If you have already set a session with us, please check that all your routes are sent. We find around a few dozen of operators, that forget one or two of their prefixes and create noise in our finding methods. So we would have a reliable data about your prefixes and in the future can automatically detect this type of hijacks for your address space. When you find out about this situation, you can also try to mitigate it. The first approach — announce routes with these more specific prefixes by yourself. In case of a new attack on these prefixes — repeat. The second approach — punish an attacker and those for whom he is a critical point (for good routes) by restricting access of your routes to him. You can do this by adding their ASN in AS\_PATH of your old routes and thus avoiding their AS by exploiting BGP route loop detection mechanism for your good, which we [described earlier](https://radar.qrator.net/blog/no-filters-shoot-your-foot).
https://habr.com/ru/post/447776/
null
null
2,734
64.51
On Tue, Jul 05, 2005 at 03:19:40PM +0100, David Vrabel wrote:> The 8250 serial driver detects the Exar XR16L2551 as a 16550A. The> XR16L2551 has an EFR register and sleep capabilities (UART_CAP_FIFO |> UART_CAP_EFR | UART_CAP_SLEEP). However, broken_efr() thinks it's a> buggy Exar ST16C255x.Grumble!> Any suggestion on how to differentiate between the two parts? Exar have> made the ST16C255x with the same registers as the XR16L255x...If they're 100% identical in terms of the registers, then it's probablyimpossible to differentiate between the two.> Perhaps it's okay for the ST16C255x to be detected as something with> UART_CAP_EFR | UART_CAP_SLEEP even if it doesn't work? i.e., by removing> broken_efr().I don't know - maybe Alex Williamson can try your patch out.> Also, the initial IER test was failing (after a soft reboot) with the> XR16L2551 part since the sleep mode bit was set but was read-only. It> seems sensible to make this test only look at the lower 4 bits.... or maybe this can be used to test for the buggy version.> Index:,The docs I've just pulled of Exar's site imply that the XR16L2551doesn't have a transmit trigger threshold, so UART_FCR_T_TRIG_00shouldn't be specified here. Also, is there a reason for restrictingthe RX trigger level to 4 instead of 8 bytes?-- Russell King Linux kernel 2.6 ARM Linux - maintainer of: 2.6 Serial core%statepending%patchIndex:,+ .flags = UART_CAP_FIFO | UART_CAP_EFR | UART_CAP_SLEEP,+ }, }; static _INLINE_ unsigned int serial_in(struct uart_8250_port *up, int offset)@@ -573,6 +581,15 @@ up->port.type = PORT_16850; return; }+ /* The Exar XR16L255x has an DVID of 0x02. This will misdetect the+ * Exar ST16C255x (A2 revision) which has registers like the XR16L255x+ * but doesn't have a working sleep mode. However, it's safe to claim+ * sleep capabilities even if they don't work. See also+ * */+ if (id2 == 0x02) {+ up->port.type = PORT_XR16550;+ return;+ } /* * It wasn't an XR16C850.@@ -585,7 +602,7 @@ */ if (size_fifo(up) == 64) up->port.type = PORT_16654;- else+ else if(size_fifo(up) == 32) up->port.type = PORT_16650V2; } @@ -611,19 +628,6 @@ up->port.type = PORT_16450; } -static int broken_efr(struct uart_8250_port *up)-{- /*- * Exar ST16C2550 "A2" devices incorrectly detect as- * having an EFR, and report an ID of 0x0201. See- *- */- if (autoconfig_read_divisor_id(up) == 0x0201 && size_fifo(up) == 16)- return 1;-- return 0;-}- /* * We know that the chip has FIFOs. Does it have an EFR? The * EFR is located in the same register position as the IIR and@@ -661,7 +665,7 @@ * (other ST16C650V2 UARTs, TI16C752A, etc) */ serial_outp(up, UART_LCR, 0xBF);- if (serial_in(up, UART_EFR) == 0 && !broken_efr(up)) {+ if (serial_in(up, UART_EFR) == 0) { DEBUG_AUTOCONF("EFRv2 "); autoconfig_has_efr(up); return;@@ -828,7 +832,7 @@ #endif scratch3 = serial_inp(up, UART_IER); serial_outp(up, UART_IER, scratch);- if (scratch2 != 0 || scratch3 != 0x0F) {+ if ((scratch2 & 0x0f) != 0 || (scratch3 & 0x0f) != 0x0F) { /* * We failed; there's nothing here */Index: linux-2.6-working/include/linux/serial_core.h===================================================================--- linux-2.6-working.orig/include/linux/serial_core.h 2005-07-04 13:43:13.000000000 +0100+++ linux-2.6-working/include/linux/serial_core.h 2005-07-05 15:07:58.000000000 +0100@@ -39,7 +39,8 @@ #define PORT_RSA 13 #define PORT_NS16550A 14 #define PORT_XSCALE 15-#define PORT_MAX_8250 15 /* max port ID */+#define PORT_XR16550 16+#define PORT_MAX_8250 16 /* max port ID */ /* * ARM specific type numbers. These are not currently guaranteed
https://lkml.org/lkml/2005/7/6/252
CC-MAIN-2021-04
refinedweb
547
59.8
Files:. def init() lcd = Qt::LCDNumber.new(2) lcd.setSegmentStyle(Qt::LCDNumber::Filled) @slider = Qt::Slider.new(Qt::Horizontal) @slider.setRange(0, 99) @slider.setValue(0) @label = Qt::Label.new() @label.setAlignment(Qt::AlignHCenter.to_i | Qt::AlignTop.to_i) connect(@slider, SIGNAL('valueChanged(int)'), lcd, SLOT('display(int)')) connect(@slider, SIGNAL('valueChanged(int)'), self, SIGNAL('valueChanged(int)')) layout = Qt::VBoxLayout.new() layout.addWidget(lcd) layout.addWidget(@slider) layout.addWidget(@label) setLayout(layout) setFocusProxy(@slider) end, which represents the time 00:00:00. Next we fetch the number of seconds from midnight until now and use it as a random seed. See the documentation for Qt::Date, Qt::Time, and Qt::DateTime for more information. Finally we calculate the target's center point. We keep it within the rectangle (x = 200, y = 35, width = 190, height = 255), i.e., the possible x and y values are 200 to 389 and 35 to 289, respectively) in a coordinate system where we put y position 0 at the bottom edge of the widget and let y values increase upwards x is as normal, with 0 at the left edge and with x values increasing to the right. By experimentation we have found this to always be in reach of the shot. point uses y coordinate 0 at the bottom of the widget. We calculate the point in widget coordinates before we call Qt::Rect::moveCenter(). The reason we have chosen this coordinate mapping is to fix the distance between the target and the bottom of the widget. Remember that the widget can be resized by the user or the program at any time.() that returns a Qt::Region so you can have really accurate collision detection. Make a moving target. Make sure that the target is always created entirely on-screen. Make sure that the widget cannot be resized so that the target isn't visible. [Hint: Qt::Widget::setMinimumSize() is your friend.] Not easy; make it possible to have several shots in the air at the same time. [Hint: Make a Shot class.]
https://techbase.kde.org/index.php?title=Development/Tutorials/Qt4_Ruby_Tutorial/Chapter_12&diff=prev&oldid=73670
CC-MAIN-2016-44
refinedweb
340
50.53
Hello, if I start to distribute some software I have written under a copyright, using a free software license, I would like to know what do the dates of the copyright stand for. If I say (c) 2009-2010 Vernon Swanepoel, what does it mean that I did on those dates. The reason I am asking is can I do something like this (written in Python) def copyrightText(): from time import localtime if localtime()[0] == 2009: crD = "2009" else: crD = "2009 - %d" % localtime()[0] return crD So that each year, the material's copyright simply updates itself. Is there a legal issue to doing that. I am not just asking Python specifically, if I do the same on a website, is there an problem with that. Basically, if I am lazy, will I get burned :S One of the main reasons I would like to do this is for websites I build for clients, so that years from now they don't have to worry that the website's copyright thingi (heavy legal terminology, I know) is updated.
https://www.daniweb.com/programming/computer-science/threads/251886/software-copyright
CC-MAIN-2017-34
refinedweb
178
67.72
This is Part I of our three-part series on video game physics. For the rest of this series, see: Part II: Collision Detection for Solid Objects Part III: Constrained Rigid Body Simulation. The following is an example of a particle simulation in the C language. #define NUM_PARTICLES 1 // Two dimensional vector. typedef struct { float x; float y; } Vector2; // Two dimensional particle. typedef struct { Vector2 position; Vector2 velocity; float mass; } Particle; // Global array of particles. Particle particles[NUM_PARTICLES]; // Prints all particles' position to the output. We could instead draw them on screen // in a more interesting application. void PrintParticles() { for (int i = 0; i < NUM_PARTICLES; ++i) { Particle *particle = &particles[i]; printf("particle[%i] (%.2f, %.2f)\n", i, particle->position.x, particle->position.y); } } // Initializes all particles with random positions, zero velocities and 1kg mass. void InitializeParticles() { for (int i = 0; i < NUM_PARTICLES; ++i) { particles[i].position = (Vector2){arc4random_uniform(50), arc4random_uniform(50)}; particles[i].velocity = (Vector2){0, 0}; particles[i].mass = 1; } } // Just applies Earth's gravity force (mass times gravity acceleration 9.81 m/s^2) to each particle. Vector2 ComputeForce(Particle *particle) { return (Vector2){0, particle->mass * -9.81}; } void RunSimulation() { float totalSimulationTime = 10; // The simulation will run for 10 seconds. float currentTime = 0; // This accumulates the time that has passed. float dt = 1; // Each step will take one second. InitializeParticles(); PrintParticles(); while (currentTime < totalSimulationTime) { // We're sleeping here to keep things simple. In real applications you'd use some // timing API to get the current time in milliseconds and compute dt in the beginning // of every iteration like this: // currentTime = GetTime() // dt = currentTime - previousTime // previousTime = currentTime sleep(dt); for (int i = 0; i < NUM_PARTICLES; ++i) { Particle *particle = &particles[i]; Vector2 force = ComputeForce(particle); Vector2 acceleration = (Vector2){force.x / particle->mass, force.y / particle->mass}; particle->velocity.x += acceleration.x * dt; particle->velocity.y += acceleration.y * dt; particle->position.x += particle->velocity.x * dt; particle->position.y += particle->velocity.y * dt; } PrintParticles(); currentTime += dt; } } If you call the RunSimulation function (for a single particle) it will print something like: particle[0] (-8.00, 57.00) particle[0] (-8.00, 47.19) particle[0] (-8.00, 27.57) particle[0] (-8.00, -1.86) particle[0] (-8.00, -41.10) particle[0] (-8.00, -90.15) particle[0] (-8.00, -149.01) particle[0] (-8.00, -217.68) particle[0] (-8.00, -296.16) particle[0] (-8.00, -384.45) particle[0] (-8.00, -482.55) As you can see, the particle started at the (-8, 57) position and then its y coordinate started to drop faster and faster, because it was accelerating downwards under the force of gravity. The following animation gives a visual representation of a sequence of three steps of a single particle being simulated: Initially, at t = 0, the particle is at p0. After a step it moves in the direction its velocity vector v0 was pointing. In the next step a force f1 is applied to it, and the velocity vector starts to change as if it was being pulled in the direction of the force vector. In the next two steps, the force vector changes direction but continues pulling the particle upwards. can neglect deformations. A rigid body is like an extension of a particle because it also has mass, position and velocity. Additionally, it has volume and shape, and so it can rotate. That adds more complexity than it sounds, especially in three dimensions. A rigid body naturally rotates around its center of mass, and the position of a rigid body is considered to be the position of its center of mass. We define the initial state of the rigid body with its center of mass at the origin and a rotation angle of zero. Its position and rotation at any instant of time t is going to be an offset of the initial state. The center of mass is the mid-point of the mass distribution of a body. If you imagine that a rigid body with mass M is made up of N tiny particles, each with mass mi and location ri inside the body, the center of mass can be computed as: This formula shows that the center of mass is the average of the particle positions weighted by their mass. If the density of the body is uniform throughout, the center of mass is the same as the geometric center of the body shape, also known as the centroid. Game physics engines usually only support uniform density, so the geometric center can be used as the center of mass. Rigid bodies are not made up of a finite number of discrete particles though, they’re continuous. For that reason, we should calculate it using an integral instead of a finite sum, like this: where r is the position vector of each point, and 𝜌 (rho) is a function that gives the density at each point within the body. Essentially, this integral does the same thing as the finite sum, however it does so in a continuous volume V. Since a rigid body can rotate, we have to introduce its angular properties, which are analogous to a particle’s linear properties. In two dimensions, a rigid body can only rotate about the axis that points out of the screen, hence we only need one scalar to represent its orientation. We usually use radians (which go from 0 to 2π for a full circle) as a unit here instead of angles (that go from 0 to 360 for a full circle), as this simplifies calculations. In order to rotate, a rigid body needs some angular velocity, which is a scalar with unit radians per second and which is commonly represented by the greek letter ω (omega). However, to gain angular velocity the body needs to receive some rotational force, which we call torque, represented by the greek letter τ (tau). Thus, Newton’s Second Law applied to rotation is: where α (alpha) is the angular acceleration and I is the moment of inertia. For rotations, the moment of inertia is analogous to mass for linear motion. It defines how hard it is to change the angular velocity of a rigid body. In two dimensions it is a scalar and is defined as: where V means that this integral should be performed for all points throughout body volume, r is the position vector of each point relative to the axis of rotation, r2 is actually the dot product of r with itself, and 𝜌 is a function that gives the density at each point within the body. For example, the moment of inertia of a 2-dimensional box with mass m, width w and height h about its centroid is: Here you can find a list of formulas to compute the moment of inertia for a bunch of shapes about different axes. When a force is applied to a point on a rigid body it, it may produce torque. In two dimensions, the torque is a scalar, and the torque τ generated by a force f applied at a point on the body which has an offset vector r from the center of mass is: where θ (theta) is the smallest angle between f and r. The previous formula is exactly the formula for the length of the cross product between r and f. Thus, in three dimensions we can do this: A two-dimensional simulation can be seen as a three-dimensional simulation where all rigid bodies are thin and flat, and it all happens on the xy-plane, meaning there’s no movement in the z-axis. This means that f and r are always in the xy plane and so τ will always have zero x and y components because the cross product will always be perpendicular to the xy-plane. This in turn means it will always be parallel to the z-axis. Thus, only the z component of the cross product matters. What this leads to is that the calculation of torque in two dimensions can be simplified to: It’s incredible how adding only one more dimension to the simulation makes things considerably more complicated. In three dimensions, the orientation of a rigid body has to be represented with a quaternion, which is a kind of four-element vector. The moment of inertia is represented by a 3x3 matrix, called the inertia tensor, which is not constant since it depends on the rigid body orientation, and thus varies over time as the body rotates. To learn all the details about 3D rigid body simulation, you can check out the excellent Rigid Body Simulation I — Unconstrained Rigid Body Dynamics, which is also part of Witkin and Baraff’s Physically Based Modeling: Principles and Practice course. The simulation algorithm is very similar to that of particle simulation. We only have to add the shape and rotational properties of the rigid bodies: #define NUM_RIGID_BODIES 1 // 2D box shape. Physics engines usually have a couple different classes of shapes // such as circles, spheres (3D), cylinders, capsules, polygons, polyhedrons (3D)... typedef struct { float width; float height; float mass; float momentOfInertia; } BoxShape; // Calculates the inertia of a box shape and stores it in the momentOfInertia variable. void CalculateBoxInertia(BoxShape *boxShape) { float m = boxShape->mass; float w = boxShape->width; float h = boxShape->height; boxShape->momentOfInertia = m * (w * w + h * h) / 12; } // Two dimensional rigid body typedef struct { Vector2 position; Vector2 linearVelocity; float angle; float angularVelocity; Vector2 force; float torque; BoxShape shape; } RigidBody; // Global array of rigid bodies. RigidBody rigidBodies[NUM_RIGID_BODIES]; // Prints the position and angle of each body on the output. // We could instead draw them on screen. void PrintRigidBodies() { for (int i = 0; i < NUM_RIGID_BODIES; ++i) { RigidBody *rigidBody = &rigidBodies[i]; printf("body[%i] p = (%.2f, %.2f), a = %.2f\n", i, rigidBody->position.x, rigidBody->position.y, rigidBody->angle); } } // Initializes rigid bodies with random positions and angles and zero linear and angular velocities. // They're all initialized with a box shape of random dimensions. void InitializeRigidBodies() { for (int i = 0; i < NUM_RIGID_BODIES; ++i) { RigidBody *rigidBody = &rigidBodies[i]; rigidBody->position = (Vector2){arc4random_uniform(50), arc4random_uniform(50)}; rigidBody->angle = arc4random_uniform(360)/360.f * M_PI * 2; rigidBody->linearVelocity = (Vector2){0, 0}; rigidBody->angularVelocity = 0; BoxShape shape; shape.mass = 10; shape.width = 1 + arc4random_uniform(2); shape.height = 1 + arc4random_uniform(2); CalculateBoxInertia(&shape); rigidBody->shape = shape; } } // Applies a force at a point in the body, inducing some torque. void ComputeForceAndTorque(RigidBody *rigidBody) { Vector2 f = (Vector2){0, 100}; rigidBody->force = f; // r is the 'arm vector' that goes from the center of mass to the point of force application Vector2 r = (Vector2){rigidBody->shape.width / 2, rigidBody->shape.height / 2}; rigidBody->torque = r.x * f.y - r.y * f.x; } void RunRigidBodySimulation() { float totalSimulationTime = 10; // The simulation will run for 10 seconds. float currentTime = 0; // This accumulates the time that has passed. float dt = 1; // Each step will take one second. InitializeRigidBodies(); PrintRigidBodies(); while (currentTime < totalSimulationTime) { sleep(dt); for (int i = 0; i < NUM_RIGID_BODIES; ++i) { RigidBody *rigidBody = &rigidBodies[i]; ComputeForceAndTorque(rigidBody); Vector2 linearAcceleration = (Vector2){rigidBody->force.x / rigidBody->shape.mass, rigidBody->force.y / rigidBody->shape.mass}; rigidBody->linearVelocity.x += linearAcceleration.x * dt; rigidBody->linearVelocity.y += linearAcceleration.y * dt; rigidBody->position.x += rigidBody->linearVelocity.x * dt; rigidBody->position.y += rigidBody->linearVelocity.y * dt; float angularAcceleration = rigidBody->torque / rigidBody->shape.momentOfInertia; rigidBody->angularVelocity += angularAcceleration * dt; rigidBody->angle += rigidBody->angularVelocity * dt; } PrintRigidBodies(); currentTime += dt; } } Calling RunRigidBodySimulation for a single rigid body outputs its position and angle, like this: body[0] p = (36.00, 12.00), a = 0.28 body[0] p = (36.00, 22.00), a = 15.28 body[0] p = (36.00, 42.00), a = 45.28 body[0] p = (36.00, 72.00), a = 90.28 body[0] p = (36.00, 112.00), a = 150.28 body[0] p = (36.00, 162.00), a = 225.28 body[0] p = (36.00, 222.00), a = 315.28 body[0] p = (36.00, 292.00), a = 420.28 body[0] p = (36.00, 372.00), a = 540.28 body[0] p = (36.00, 462.00), a = 675.28 body[0] p = (36.00, 562.00), a = 825.28 Since an upward force of 100 Newtons is being applied, only the y coordinate changes. The force is not being applied directly on the center of mass, and thus it induces torque, causing the rotational angle to increase (counter clockwise rotation). The graphical result is similar to the particles animation but now we have shapes moving and rotating. This is an unconstrained simulation, because the bodies can move freely and they do not interact with each other – they don’t collide. Simulations without collisions are pretty boring and not very useful. In the next installment of this series we’re gonna talk about collision detection, and how to solve a collision once it’s been detected. Appendix: Vectors A vector is an entity that has a length (or, more formally, a magnitude) and a direction. It can be intuitively represented in the Cartesian coordinate system, where we decompose it into two components, x and y, in a 2-dimensional space, or three components, x, y, and z, in 3-dimensional space. We represent a vector with a bold lowercase character, and its components by a regular identical character with the component subscript. For example: represents a 2-dimensional vector. Each component is the distance from the origin in the corresponding coordinate axis.. Some of the fundamental algebraic operations that operate on vectors are reviewed below. Length The length, or magnitude, operator is represented by || ||. The length of a vector can be computed from its components using the famous Pythagorean theorem for right triangles. For example, in two dimesions: Negation When a vector is negated, the length remains the same, but the direction changes to the exact opposite. For example, given: the negation is: Addition and Subtraction Vectors can be added to, or subtracted from, one another. Subtracting two vectors is the same as adding one to the negation of the other. In these operations, we simply add or subtract each component: The resulting vector can be visualized as pointing to the same point that the two original vectors would if they were connected end-to-end: Scalar Multiplication When a vector is multiplied by a scalar (for the purposes of this tutorial, a scalar is simply any real number), the vector’s length changes by that amount. If the scalar is negative, it also makes the vector point in the opposite direction. Dot Product The dot product operator takes two vectors and outputs a scalar number. It is defined as: where ø is the angle between the two vectors. Computationally, it is much simpler to compute it using the components of the vectors, i.e.: The value of the dot product is equivalent to the length of the projection of the vector a onto the vector b, multiplied by the length of b. This action can also be flipped, taking the length of the projection of b onto a, and multiplying by the length of a, producing the same result. This means that the dot product is commutative – the dot product of a with b is the same as the dot product of b with a. One useful property of the dot product is that its value is zero if the vectors are orthogonal (the angle between them is 90 degrees), because cos 90 = 0. Cross Product In three dimensions, we can also multiply two vectors to output another vector using the cross product operator. It is defined as: and the length of the cross product is: where ø is the smallest angle between a and b. The result is a vector c that is orthogonal (perpendicular) to both a and b. If a and b are parallel, that is, the angle between them is zero or 180 degrees, c is going to be the zero vector, because sin 0 = sin 180 = 0. Unlike the dot product, the cross product is not commutative. The order of the elements in the cross product is important, because: The direction of the result can be determined by the right hand rule. Open your right hand and point your index finger in the same direction of a and orient your hand in a way that when making a fist your fingers move straight towards b through the smallest angle between these vectors. Now your thumb is pointing in the direction of a × b.
https://www.toptal.com/game/video-game-physics-part-i-an-introduction-to-rigid-body-dynamics
CC-MAIN-2017-26
refinedweb
2,734
55.74
This is the mail archive of the gcc@gcc.gnu.org mailing list for the GCC project. -B vs Multilib Re: -Wparentheses lumps too much together 4.3.0 manual vs changes.html 4.3.0-rc2 available 4.3.1 Status Report (2008-03-31) [4.3/4.4]: PATCH: PR target/35453: nmmintrin.h defines macros SIDD_XXX [Ada] GNAT BUG Assert_Failure uintp.adb:1593 Re: [Ada] gnat.dg/socket[12].adb test duplicates? Re: [Ada] no run-time compilation (was: Clean up No_Run_Time tests in exp_ch4.adb) Re: [linux-cirrus] Re: wot to do with the Maverick Crunch patches? Re: [linux-cirrus] wot to do with the Maverick Crunch patches? [PATCH] BIT_FIELD_REF_UNSIGNED considered harmful [PATCH] Fix bug on soft emulation of float point in gcc/longlong.c Re: [PATCH] linux/fs.h - Convert debug functions declared inline __attribute__((format (printf,x,y) to statement expression macros Re: [PATCH] linux/fs.h - Convert debug functions declared inline __attribute__((format (printf,x,y) to statement expression macros Re: [PATCH] Make alias_sets_conflict_p less conservative Re: [PATCH][4.3] Deprecate -ftrapv [PATCH][RFC] Statistics "infrastructure" [PING] Issues stopping us from removing libcall notes [RFA] optimizing predictable branches on x86 [RFC] GCC caret diagnostics [RFH] Uninitialized warning as error is disabled on the trunk Re: [trunk] Addition to subreg section of rtl.text. [tuples] Branch frozen (again) Re: [tuples] Branch frozen (again) [tuples] gimple_assign_subcode for GIMPLE_SINGLE_RHS [tuples] libgfortran building is broken [tuples] Merged trunk->branch @132948 [tuples] Tuples branch frozen [wwwdocs] Re: GCC 4.2.4 Status Report (2008-03-15) [wwwdocs] Re: GCC 4.3.1 Status Report (2008-03-15) [wwwdocs] Re: GCC 4.4 schedule ABI64 and register save area Adding new section attribute aneurin chandras hammond griswold Re: API for callgraph and IPA passes for whole program optimization ARM/Thumb function attribute atomic accesses Auto-vectorization: need to know what to expect Basic block infrastructure after dbr pass Re: Been Looking how gcc operates there is a major weaknesses in its optimiser. Re: Been Looking how gcc operates there is a major weaknesses in its optimiser. Been Looking how gcc operates there is a major weaknesses in its optimiser. Re: Benchmarks: 7z, bzip2 & gzip. Re: birthpoints in rtl. Bootstrap failure building libobjc fixed bootstrap failure on powerpc-linux C++ FE question: When is CLASSTYPE_VBASECLASSES valid? C++0x rvalue references get clobbered somehow cc1plus-dummy linkage Combine repeats matching on insn pairs and will ICE on 3. Combined gcc + binutils source tree doesn't bootstrap Come join me on ssbbw4u... Company Offer compiler slowdown in 4.3 development Constrain valid arguments to BIT_FIELD_REF Copy constructor access check while initializing a reference Could someone please check if FSF received papers for Intel engineers? Re: Could someone please check if FSF received papers for Intel engineers? CRX and CR16 port maintainer DFA state and arc explosion dg-skip-if was Re: gcc 4.3.0 i386 default question Different *CFLAGS in gcc/Makefile.in Re: Draft SH uClinux FDPIC ABI E-gold owner Douglas Jackson has been killed The effects of closed loop SSA and Scalar Evolution Const Prop. Enum in namespace An error occured when building gcc4.3.0 Exception handling on AIX5.3 with gcc 3.4.6 Re: Excess registers pushed - regs_ever_live not right way? executable stack in gcc shared libs? Forward propagation before register allocation gcc 3.3.2 Wind River VxWorks PowerPC gcc 4.2.3 : make: *** [bootstrap] Error 2 GCC 4.2.4 Status Report (2008-03-15) GCC 4.2.4 Status Report (2008-03-31) GCC 4.3 - Error: Link tests are not allowed after GCC_NO_EXECUTABLES GCC 4.3 license in manual still under GPLv2 GCC 4.3.0 compilation error gcc 4.3.0 i386 default question GCC 4.3.0 Status Report (2008-03-03) GCC 4.3.0 Status Report (2008-03-06) GCC 4.3.0 Status Report (2008-03-06, 2nd editionOF) GCC 4.3.0-20080228 (powerpc-linux-gnuspe) ICE on 20000718.c test GCC 4.3.1 Status Report (2008-03-15) GCC 4.4 schedule GCC : how to add VFPU to PSP Allegrex (MIPS target) ? GCC build problem GCC extension proposal (loop unswitching) gcc for Linux/x86 vs. Solaris/x86 gcc participation in C standards process Gcc support for Intel AVX Re: GCC wiki very slow? gcc-4.1-20080303 is now available gcc-4.1-20080310 is now available gcc-4.1-20080317 is now available gcc-4.1-20080324 is now available gcc-4.1-20080331 is now available gcc-4.2-20080305 is now available gcc-4.2-20080312 is now available gcc-4.2-20080319 is now available gcc-4.2-20080326 is now available gcc-4.3-20080306 is now available gcc-4.3-20080313 is now available gcc-4.3-20080320 is now available gcc-4.3-20080327 is now available gcc-4.3.0/ppc32 inline assembly produces bad code gcc-4.4-20080307 is now available gcc-4.4-20080314 is now available gcc-4.4-20080321 is now available gcc-4.4-20080328 is now available gcj broken on darwin genattrtab segfault on RH 7.3 (powerpc cross) Getting GCC to always dllimport vtables on X86? gimple GNAT on SVN Trunk GNU linker ld Google Summer of Code 2008 Google Summer of Code 2008 mentors GSoC - Multithreaded linker GSoC ideas GSOC Student application Have proposals for 2008 gcc summit been reviewed? hc11/12 port extension to s12x Help with GCC on Cygwin How should _Decimal64 and _Decimal128 be aligned on stack? How to understand gcc source code? howto run cross testings with help of translators needs a bit of help RE: Idea for code size reduction Implementing a restrictive addressing mode for a gcc port Information regarding issue with While Loop with O3 optimization Injecting data declarations? insn appears multiple times Re: Interoperability of Fortran array and C vector? Invalid address after reload Is vec_init<mode> allowed to FAIL? Re: Issues stopping us from removing libcall notes Re: larger default page sizes... The Latest Lexer/cpplib improvements libtool for shared objects? The Linux binutils 2.18.50.0.5 is released Linux doesn't follow x86/x86-64 ABI wrt direction flag Lots of new test failures on trunk due to unrecognizable insn Re: MAINTAINERS update (was: [wwwdocs] java/projects.html -- remove broken link) Re: MAINTAINERS update (was: [wwwdocs] java/projects.html -- remove broken link) MAINTAINERS update (was: [wwwdocs] java/projects.html -- remove broken link) minor mistake in = SSSE3 --> SSE3 Re: minor mistake in = SSSE3 --> SSE3 Need link ln -s ../lib64/libgomp.spec libgomp.spec in $prefix/lib on x86_64-unknown-linux-gnu New picoChip port and maintainers new regression on 4.3.0 branch New wiki page for openmp and parallelization Obvious problem with Ada front-end patches handling Official GCC git repository Re: optimizing predictable branches on x86 parallel bootstrap failure: options.h Re: Patch: delete treelang please add DFP to gcc-4.3/changes.html Re: plugin includes for MELT Porting gcc Fwd: Porting gcc to a new architecture Possible GCC 4.3 driver regression caused by your patch Possible gcc-4.3 regression wrt bootstrapping the toolchain Re: PR35401 and PR30572 are gcc 4.3.0 release blockers on darwin Problems with builtin setjmp receiver getting eliminated - Help RE: A proposal to align GCC stack push_secondary_reload: give me a break... Question: How can I upload my work? Regenerating configure scripts Re: Regression with ltrans-7.f90 RELEASE BLOCKER: Linux doesn't follow x86/x86-64 ABI wrt direction flag Release planning for GCC 4.4? RFC: adding knowledge of the int_fastN_t types to the compiler RFC: Idea for code size reduction RFH: testers for ieee test RTL definition Seg fault in call_gmon_start Soliciting help upgrading our bugzilla install Some regression numbers Re: SPEC / testsuite results for disabling SFTs and the alias-oracle patches SPEC / testsuite results for disabling SFTs and the alias-oracle patches SSA Vs unSSA Stan Shebs appointed maintainer for Darwin, Objc and Objc++ start.encap target in gcc/Makefile.in Re: static array with constant size Status of Mercurial mirror Subject gcc testsuite testcase gcc.c-torture/compile/20010327-1.c subreg question successful build Successful native mingw build gcc 4.3.0 RE: Successfull build of gcc 4.2.3 with MinGW 5.13 in windows XP Swing replacements Test Harness and SPARC VIS Instructions Re: testsuite run-time test for sse2 support Thread starvation and resource saturation in atomicity functions? Un-deprecating CRX: Request to review and commit Unaligned attribute update_path in gcc/prefix.h? va_list bug? valgrind "uninitialized value" for __cxa_begin_catch? (Only if -static) vectorizer default in 4.3.0 changes document missing What licence(s) does GNAT and the runtime fall under now? why don't have 'umul' rtx code wot to do with the Maverick Crunch patches? XDELETEVEC xscale-elf-gcc: compilation of header file requested
http://gcc.gnu.org/ml/gcc/2008-03/subjects.html#00258
crawl-003
refinedweb
1,481
60.31