A Premium Guide to Editing The This Can Be Used As A Simple Cover Letter Insert
Below you can get an idea about how to edit and complete a This Can Be Used As A Simple Cover Letter Insert step by step. Get started now.
- Push the“Get Form” Button below . Here you would be transferred into a splasher allowing you to make edits on the document.
- Choose a tool you require from the toolbar that pops up in the dashboard.
- After editing, double check and press the button Download.
- Don't hesistate to contact us via [email protected] if you need further assistance.
The Most Powerful Tool to Edit and Complete The This Can Be Used As A Simple Cover Letter Insert


A Simple Manual to Edit This Can Be Used As A Simple Cover Letter Insert Online
Are you seeking to edit forms online? CocoDoc is ready to give a helping hand with its useful PDF toolset. You can utilize it simply by opening any web brower. The whole process is easy and quick. Check below to find out
- go to the CocoDoc's free online PDF editing page.
- Upload a document you want to edit by clicking Choose File or simply dragging or dropping.
- Conduct the desired edits on your document with the toolbar on the top of the dashboard.
- Download the file once it is finalized .
Steps in Editing This Can Be Used As A Simple Cover Letter Insert on Windows
It's to find a default application that can help make edits to a PDF document. Fortunately CocoDoc has come to your rescue. Examine the Manual below to find out possible approaches to edit PDF on your Windows system.
- Begin by obtaining CocoDoc application into your PC.
- Upload your PDF in the dashboard and make modifications on it with the toolbar listed above
- After double checking, download or save the document.
- There area also many other methods to edit your PDF for free, you can check this guide
A Premium Handbook in Editing a This Can Be Used As A Simple Cover Letter Insert on Mac
Thinking about how to edit PDF documents with your Mac? CocoDoc can help.. It empowers you to edit documents in multiple ways. Get started now
- Install CocoDoc onto your Mac device or go to the CocoDoc website with a Mac browser. Select PDF paper from your Mac device. You can do so by hitting the tab Choose File, or by dropping or dragging. Edit the PDF document in the new dashboard which includes a full set of PDF tools. Save the file by downloading.
A Complete Instructions in Editing This Can Be Used As A Simple Cover Letter Insert on G Suite
Intergating G Suite with PDF services is marvellous progess in technology, with the power to streamline your PDF editing process, making it quicker and more cost-effective. Make use of CocoDoc's G Suite integration now.
Editing PDF on G Suite is as easy as it can be
- Visit Google WorkPlace Marketplace and get CocoDoc
- install the CocoDoc add-on into your Google account. Now you are in a good position to edit documents.
- Select a file desired by pressing the tab Choose File and start editing.
- After making all necessary edits, download it into your device.
PDF Editor FAQ
Do you use a form cover letter or do you add sentences in relating to the job?
Please take this advice seriously: do not ever use a one-size-fits-all or Mad Libs style cover letter.Those who’ve done a bit of hiring can tell from your cover letter when you’re spamming out a load of resumes with an unremarkable, vague and formulaic cover. The message it sends is pretty clear: this is someone who needs a job. We like people who want the job we’re offering.How do we know when you’re a serious candidate?When you write a cover letter that tells us:You invested time and effort to write a compelling argument that explains why you are uniquely qualified for the positionYou know what our business does and what the position entailsSpecific examples of your accomplishments in our field, and added texture beyond what we see on your resumeDetails that convey your seniority, professionalism, personality and fit with our cultureThat you proof-read, spell-checked and wrote with the cadence of a serious professionalHow do we know that you’re probably not among the 10 in 100 who we’ll include in our pool of prospects?When your cover letter is cut-and-pasted from one of the “Ten Cover Letters Guaranteed To Get You An Interview” or any other template from the first 20 Google search results for “Examples of good [insert industry] cover letters”. I see these all the time, and they often include grammatical and formatting errors from the source article.A cover letter that only vaguely refers to the open position: ”I’m interested in the job you posted on LinkedIn” doesn’t cut it. If we misplace your application or forward it to other decision-makers, the cover on your resume need to include the position description and reference where you learned about it. We often forward resumes internally for comments on your suitability: if our people don’t know which job you’re interested in, they can’t advocate on your behalf. The candidates who our people tell us “This is someone I’d like us to interview” or “I like this one” are the people most likely to get called first.Useless boilerplate that tells us nothing: You’re a hard worker? So are the rest of the applicants. You’re detail-oriented? If that’s true, you would include the details that are oriented to our requirements. Oh, you’re the one who’s a team player? We’re a team of professional niche consultants. You could be on a bowling team, and unless you elaborate we’re not going to find this compelling.Typos, poor grammar, text-message lingo or emojis: A letter that looks like something a teenager would post on Instagram is not someone we can trust to represent us when communicating with clients and partners.Absence of a voice in your writing, or the wrong voice: I see cover letters for senior positions and roles that are communications-oriented that read like an elementary school book report: “This is an interesting job. Somethings I’m good at are why you should interview me. Consulting is an interesting and fun job I think I’d do good at.”Facts clearly plucked out of Wikipedia: This is almost worse than typographical errors. I have dozens of cover letters that include hilariously bad laundry-lists culled from Wikipedia. “Consulting is the fourth-largest job category by revenue-per-employee. Examples of consulting companies include Deloitte, with 100 locations including Hong Kong and India. Boutique and specialty firms have an average billable rate of $150 per hour.”First of all, thanks for the update. Secondly, there are three aspects of consulting that no reputable firm discusses publicly: Our client roster, our financials and the rates we bill for the services we provide. If you’re experienced in what we do, you know this. And if we have even a whiff of concern that a prospective hire lacks the discretion, sensitivity and appreciation for what our clients expect from us, we’ll err on the side of caution and go with someone who knows the metaphorical secret handshakes that tell us they’re the real deal.Remember that most employers read submissions from prospects from a positive starting point:Is this the person who we need and is better than the rest?When a candidate is removed from consideration, it’s almost always the result of the candidate disqualifying themselves. We don’t begin the hiring process by trying to find reasons not to hire people… You start from a position with an automatic 50% chance we’ll want to hire you. Each positive you provide moves you to 60%, 80% and eventually to the final 5% who are contenders.Your cover letter can either build on the free 50% we automatically credit to your application, or it can chip away 5% for each unforced error and 10% for each glaring mistake you make introducing yourself. One failure to provide the minimum expectations puts you at a disadvantage right away. When a cover letter produces three or four strikes against you, that puts you in the bottom-fifth of our candidate pool before we even see your resume.Do the obvious work in writing your cover letter, and your resume will be read with our belief that you have 70% to more than 90% of what we need from a candidate.Think about who’s more likely to get the job: Someone who earns our interest immediately? Or someone who self-inflicts damage on their prospects before we turn the first page? It’s really that simple.Good luck, and thanks for the A2A!
What's the best way to get your résumé noticed?
You want your resume to stand out.You know that when you apply for that job online, you’re competing with 100s of other applicants.Your resume is a needle in a haystack. What’s going to make it so special that it screams “Read Me” to the person receiving it at the other end?The thought of creating a resume that looks different appeals to you at first.You might have checked Google Images or Pinterest for “Good Resume Designs” and was amazed by those colorful resumes with the text blocks and shapes and dots and pictures.If you choose to create a resume that has got a design full of color and flair, you are headed for rejection city.Are you confused?How can a document so attractive and colorful not get noticed?There’s a simple explanation.First, I’m going to go old school.And at the end, I’ll tell you about the 21st century resume that gets noticed.The playersFirst, lets look at the journey your resume takes so you understand why a Las Vegas resume will not work.There are 3 “people” looking at your resume.#1 A software called an application tracking system (ATS)The purpose of this software is to make the recruiter’s life easier. To filter out all the candidates that it “thinks” don’t qualify for the job and leave behind the ones that do for the recruiter to look at.The ATS can be a private system invested by the company, typically used if you have to apply for the job at the company’s website.Or a third party system that you’ve come to know and love - The LinkedIns, Monsters, Indeeds that you see online.#2 A recruiterThe purpose of the recruiter is to make the hiring manager’s job easier.They may now filter out all the candidates the ATS failed to weed out, by scanning each resume for about 6 seconds and short listing the ones they think the hiring manager will like, and declining the ones they think won’t make the cut(The recruiter also filters out applicants using a more human approach - by actually speaking to candidates over a phone conversation.)#3 The hiring managerThe boss. The alpha dog. The decider of your career and income for the years to come. The one person who has the power (and budget) to give you that job.This is also the person who is on a tight schedule and the hiring process eats into their busy schedule, so they want to make sure they are spending their time with worthy candidates.After the ATS and Recruiters have filtered the 100s of applications that came down the pipeline, the handful of residual resumes that made the cut is what finally gets viewed by the person with the authority to give you the future you deserve.The approachAssuming you are applying for a job in a medium sized firm or larger, your resume is going to go through all three phases above.Logically, you’d want your resume to cut the line and drop straight into the hiring manager’s lap right?While that is possible (and we’ll get to that later on), if you are (or forced to) follow the typical application path, your resume has to past the first two tests:#1 Keep that pesky ATS “happy”#2 Catch the recruiter’s attention in that short 6 second window you haveThe best way to get your resume noticedHere it is:KEEP IT SIMPLE. KEEP IT RELEVANT.Why simple?Let’s talk about the ATS for a second.What do all machines have in common?They like to keep things boring.If you could ask a digital machine how they would like to receive information, they would probably tell you to give it to them like this:10100001110101011010100101000101101111001001001001011110101100101110101101011010An ATS is a digital software that is scanning your resume for text.The more graphics you have in there, and the more you have your text in boxes and columns and other fancy magazine-like layouts, the more you are going to confuse it.And a confused resume is a rejected resume.Now, on to the recruiter.You might be thinking, finally a human being is going to look at my resume. Some of that fancy designs I have worked so hard on is going to help.Wrong.Remember those 6 seconds?When recruiters are going through those resumes, they are in what I like to call assembly-line mode.They are recursively scanning those resumes for the keywords that were discussed between them and the hiring manager when this position was created.Recruiters are not experts in your field (compared to the hiring manager).All they know are those keywords that the hiring manager told them - “Six Sigma Green Belt” | “PMP Certified” | “10+ years experience” | “Automotive Industry experience preferred” | “French Speaking” | “Bachelors in Computer Science” - and these words resonate in their heads during those 6 seconds.If you don’t make it easy for them to catch these words in those 6 seconds - [DECLINED!]Just like an assembly line at a factory, if a resume coming down that assembly belt is out of the ordinary, it gets rejected.So, my dear job seeker, as much as an attractive resume appeals to you, and may impress your family and friends, it’s not going to work in real life.(The only exception to this is if you are applying for a job that requires creative design)Why relevant?So now that you know how a simple design will make your resume readable to an ATS and recruiter, let’s talk relevance.Once your resume is being read it must be short listed.It must contain the keywords that is relevant to the job you are applying for.I know.That stupid jargon “Keyword” again.Why does everyone keep talking about it and how do I know what they are for my industry?Remember our old friend the ATS? Another thing machines have in common, they need to be provided logic to work.The logic an ATS has to filter out resumes is by providing it keywords.As stated earlier, these keywords are discussed between the hiring manager and recruiter. They create the job description from it and feed it to the ATS.Keywords is one of many logical instructions (Algorithms) that an ATS uses.So how do you find out which keywords to target in your resume?You can use job descriptions from online jobs boards and some clever online tools called JobScan (that was written with ATS Algorithms in mind) to get your keywords.With the relevant keywords now inserted in your simple resume, it passed the first two tests with the ATS and recruiter and landed in the lap of the hiring manager.Cutting the lineYou asked for the “best” way to get your resume noticed so I explained it considering the most conventional process that people follow when applying for jobs - submitting an application online.But I would feel guilty if I did not expose how I personally had success of getting my resume noticed using some unconventional tricks that helped my cut the line ahead of my competition.During my recent job search campaign, the final step in my process was to post my resume to the hiring manager, recruiter and VP.No, not online post, not a LinkedIn post, and not a Facebook post.Think old school people - actually printing my resume, putting it in an envelope, licking a stamp, and dropping it in a red box in the middle of -20 degrees Canadian winter weather.One recruiter actually called me up because she felt I really wanted this job because I went through so much trouble.Another approach I took was to contact recruiters and hiring managers directly from LinkedIn, get their email address, and email my resume directly to their inbox.That’s the only reason I got into the company I’m working for today.I know because I asked the recruiter who hired me why he selected me.The 21st century resumeFor my last job, I still stuck to a simple relevant resume, but I took it one step further with an online resume.I advertised my resume - connelvalentine.com - on my paper resume, cover letter and thank you note.This wasn’t a substitute for my traditional resume, but it certainly helped me get noticed.Hiring managers ultimately give you the job, and they want to make sure they are hiring people with the right personality and the right cultural fit.If you felt that a fancy designed resume will get you noticed, it won’t.Because hiring managers care more about what you can do over what you can design on paper. The smart ones know that these days you can get a fancy designed resume using online tools as well.Relevant content displayed in a simple format explaining how you will solve the hiring managers problems, how you will make him/her look good, and proof that the responsibilities you had actually made a difference will get you noticed.If you want to know how I managed to get my resumé noticed by the top companies in the country, I shared my resumé template along with a short video showing how.Click here to download the resume template-Reach for your potential-
Which are the top 20 papers related to AI (both machine learning & symbolic), so that I can cover the basics and choose a niche for my research?
Ah, the sort of challenging question that I like to ponder about on an otherwise lazy Saturday morning in the San Francisco Bay Area! I began my career in AI as a young Master’s student in Indian Institute of Technology, Kanpur, actually enrolled as a EE major, but got enchanted by Hofstadter’s Godel, Escher and Bach book into studying AI. That was 1982, so I’ve been working in AI and ML for the past 36 years. Along the way, I’ve read, oh, easily about 10,000 papers or so, give or take a few hundred. So, among these thousands of papers, now, I have to pick the “top 20 papers”, so that you, the interested Quora reader, can get a glimpse of what attracts someone like me to give up everything in pursuit of this possibly idealized quest to make machines as smart as humans and other animals. Now, there’s a challenge I can't resist.OK, any list like this is going to be 1) hopelessly biased by my personal choices 2) not entirely representative of modern AI. 3) a VERY long read! Remember that a lot of us who got into AI in the late 1970s or early 80s did so far before there was any commercial hope that AI would pay off. We were drawn to the scientific quest underlying AI: how to build a theory that explain how the brain works, how the mind was the result of the brain etc. None of us had any clue, it is safe to say, that in the early 21st century, AI would become a hugely profitable venture.But, I’m going to argue that now more than ever, it is vitally important for those entering the AI field to understand 1) where did the ideas for AI come from 2) that insights into the brain come from many fields, from neuroscience and biology to psychology and economics and from mathematics as well, so my choice of papers reflects that, and I’ve chosen papers from multiple academic fields. I’ve also not shied away from papers that are critical of things that you might believe in deeply (e.g., the power of statistical machine learning to solve potentially any AI problem).I’ll try to pepper the list with my chatty commentary as well, so it’s not going to be one of those all too boring “here’s 20 things you should know about blah”, which is all too often what you see on the web. But, with my commentary, this is going to be a really long reply. What I want to give you a glimpse of is the panoply of fascinating characters who made up this interdisciplinary quest to understand brains from a scientific and computational point of view, how diverse their backgrounds were, and what an amazingly accomplished set of minds they were. It is to their credit that AI has come along as quickly as it has, barely 60 years since it began. Without such a dazzling collection of minds working on the problem, we would have probably taken much longer to make any real progress.The list is somewhat historical and arranged chronologically as far as it is possible. I’ve also tried to keep in mind that the point of this list is that it should be comprehensible to a newbie entering the field of AI, so much as I like to include some heavy hitting math papers (of which I’ve selected a couple), I’ve included a few very sophisticated highly technical papers, since you need to get a sense of what AI is in the 21st century. So, the readability of this top 20 list varies widely: some papers are easy to get through in a Sunday afternoon. Others — well, let’s say that you’ll need several weeks of concentrated reading to make headway, assuming you have the math background. But, there are not many of the latter, so don't worry about not having the right background (yet).Let’s begin, as they say, at the beginning…..A logical calculus of the ideas immanent in nervous activity, by Warren McCullough and Walter Pitts (Univ. of Chicago), vol. 5, pp. 115–133, 1943. (http://www.cs.cmu.edu/~./epxing/Class/10715/reading/McCulloch.and.Pitts.pdf). This is the first great paper of modern computational neuroscience, written by two brilliant researchers, one senior and distinguished (McCullough), the other, Pitts, a dazzling prodigy who had had no education of any sort, but talked his way into a position with McCullough. Pitts grew up in inner city Detroit, and because he was mercilessly beaten up by gang members who were older than him, he took refuge in the Detroit Public Library. It is rumored he devoured all 1000+ pages of Bertrand Russell and Alfred North Whitehead’s Principia Mathematica in one marathon reading session of several days and nights. This is not easy reading — it is a dense logical summary of much of modern math. Pitts was brave enough, even through he was barely in high school and had had no education, to audaciously write to Bertrand Russell in England, then a famous literary figure would go on to win the Nobel Prize in literature, as well as a great mathematician, pointing out a few errors and typos in the magnum opus. This young boy so impressed Russell that later he wrote him a glowing recommendation to work with Warren McCullough. Thus was born a great collaboration, and both moved shortly to MIT, where they came under the influence of none other than Norbert Wiener, wunderkind mathematician who invented the term “cybernetics” (the study of AI in man and machines). McCullough was a larger than life character who worked all night, and seemingly subsisted on diet of “Irish whiskey and ice cream”. Pitts wrote a dazzlingly beautiful PhD thesis on “three-dimensional neural nets”, and then, as a tragic Italian opera would have it, everything fell apart. Wiener and McCullough had a falling out (so petty was the reason that I will not repeat it here), and thus, McCullough stopped being actively working with Pitt, and Pitt sort of just faded away, but sadly, not before burning the only copy of his unpublished PhD dissertation before he defended it. No copy has yet been found of this work. Read the tragic story here — warning: keep a box of Kleenex handy for at the end, you will cry! — The Man Who Tried to Redeem the World with Logic - Issue 21: Information - Nautilus (read also the classic paper “What the Frog’s Eye tells the Frog’s Brain” https://hearingbrain.org/docs/letvin_ieee_1959.pdf by the same duo. A great modern paper is the recent breakthrough in biology at Caltech where the facial code used in primate brains to identify faces has finally been cracked — The Code for Facial Identity in the Primate Brain. — showing that in one narrow area, we may know what the human eye is telling the human brain, almost 60 years after McCullough and Pitts asked the question).Steps towards Artificial Intelligence, Marvin Minsky, Proceedings of the IRE, January 1960 (http://worrydream.com/refs/Minsky%20-%20Steps%20Toward%20Artificial%20Intelligence.pdf). Many date the beginning of AI formally to this article, which really outlined the division of AI into different subfields, many of which are still around, so this paper really can be said to have been the first to lay out the modern field of AI in its current guise. Minksy was a prodigy who did a PhD in math at Princeton (like many others in AI currently and in the past), and after a dazzling postdoc at Harvard as a Fellow (where he did early work in robotics), started the highly influential MIT AI Lab, which he presided over for a number of decades. He was a larger than life character, and those who knew him well had a large stock of stories about him. Among the best I’ve heard is one where he was interviewing a faculty candidate — a rather nervous young PhD who was excitedly explaining his work on the blackboard — when the student turned around, he discovered he was alone in his office. Minksy had disappeared during his explanation. The student was mortified, but Minksy later explained that what the student had told him sounded so interesting that Minksy had to step outside and take a walk to think the ideas over. Minsky was a polymath, at home in theoretical computer science where wrote some influential papers and a book, in psychology where he was an avid disciple of Freud and wrote a paper on AI and jokes and what it meant about the subconscious, in education where he pioneered new educational learning technology, and many other fields.Programs with Common Sense, John McCarthy, in Minsky, Semantic Information Processing, pp. 403–418, 1968. (http://www-formal.stanford.edu/jmc/mcc59.ps) McCarthy was the other principal founder of AI, who after a short period of working at MIT, left to found the Stanford AI Lab, which in due course proved to be just as influential as its East coast cousin. McCarthy above all was a strong believer in the power of knowledge, and in the need for formal representations of knowledge. In this influential paper, he articulates his ideas for a software system called “an Advice Taker”, which can be instructed to do a task using hints. The Advice Taker is also endowed with common sense, and can deduce obvious conclusions from the advice given to it. For example, a self-driving car can be given the official rules of the road, as well as some advice about how humans drive (such as “in general, humans do not follow the speed limit on most highways, but tend to drive 5–10 miles above the speed limit”). Critical to McCarthy’s conceptualization, it would not be sufficient to have a neural net learn the driving task. Knowledge had to represented explicitly so it could be reasoned about. He says something profound in the paper, which may shock most modern ML researchers. He says in page 4 in italics (emphasis) that “In order for a program to be capable of learning something it must first be capable of being told it”! By this definition, McCarthy would not view most deep learning systems as really doing “learning” (for none of the deep learning systems can be told what they learn). McCarthy was also famous for his work on lambda calculus, inventing the programming language LISP, using which much of AI research was then carried out. Most of my early research in AI was done using LISP, including my first (and most highly cited paper) work on using reinforcement learning to teach robots in the early 1990s at IBM.Why should Machines Learn, by Herbert Simon, in Michalski, Carbonell, and Mitchell (editors), Machine Learning, 1983 (http://digitalcollections.library.cmu.edu/awweb/awarchive?type=file&item=33928). Herb Simon was a Nobel laureate in Economics, who spent his entire academic career at Carnegie Institute of Technology (later Carnegie Mellon University), doing much to build the luster and prestige of this now world-class university. He was one of the true polymaths, at home in half a dozen departments, from computer science to economics to business administration and psychology, in all of which he made foundational contributions. He was a gifted speaker, and I was particularly fortunate to be able to attend several presentations by Simon during the mid 1980s when I spent several years at CMU. In this article, Simon asks the question that very few AI researchers today bother asking: why should machines learn? According to Simon, why should machines, which can be programmed, bother with this slow and tedious form of knowledge acquisition, when something far quicker and more reliable is available. You’ll have to read the article to find his answers, but this article is valuable for giving perhaps the first scientific definition of the field of machine learning, a definition that is still valid today. Simon made many other contributions to AI, including his decades long collaboration with Allan Newell, another AI genius at CMU, whose singular ability in asking the right questions, made him a truly gifted researcher. It is rumored that computer chess came to life when Allan Newell mentioned casually in a conversation in the CMU CS common room about how the branching factor of chess is not all difficult to emulate in hardware, a comment that Hans Berliner followed up on in bringing the first modern chess player, Deep Thought, to fruition (the same CMU team went to IBM, built Deep Blue which of course beat Kasparov).Non-cooperative games, PhD thesis, John Nash, Princeton. (Non-Cooperative Games) John Nash came to Princeton as a 20 year old mathematician from Carnegie Institute of Technology in 1948 with a one line recommendation letter: “This man is a genius”. His PhD thesis would fully affirm his alma mater’s assessment of his capabilities. Nash took the work of von Neumann and Morgenstern’s zero-sum games into a whole new level with his dazzling generalization, which would earn him a Nobel prize decades later. Most of Nash’s history has been recounted in Sylvia Nasar’s wonderful biography A Beautiful Mind (later made into a movie starring Russell Crowe as John Nash). Legend had it that von Neumann himself did not think much of Nash’s work, calling it “another fixed point theorem”. Nash finished his ground breaking thesis in less than a year from start to finish. He arrived in Princeton in September 1948, and in November 1949, Solomon Lefschetz, a distinguished mathematician communicated the results of Nash’s thesis to the National Academy of Sciences meeting. Today, billions of dollars of product (from wireless cellular bandwidths to oil prospects) are traded using Nash’s ideas of game theory. The most influential model in deep learning today is the Generative Adversarial Network (GAN), and the key question being studied for GANs is whether and when do they converge to a Nash equilibrium. So, 70 years after Nash defended his short but Nobel prize winning thesis at Princeton, his work is still having a huge impact in ML and AI. Nash’s work also become a widely used framework to study evolutionary dynamics, giving rise to a new field called evolutionary game theory, pioneer by John Maynard Smith. Game theory is a crucial area for not only AI but also for CS. It has been said that the “Internet is just a game. We have to find what the equilibrium solution is”. Algorithmic game theory is a burgeoning area of research, studying things like “The Price of Anarchy”, or how solutions to hard optimization problems can be solved by letting millions of agents make locally selfish decisions. Nash’s PhD advisor at Princeton was Tucker, who Nash called “The Machine”. His second reader of his PhD thesis was Turkey, who can be called one of the fathers of modern machine learning, since he invented exploratory data analysis at Princeton (and later also invented the Fast Fourier Transform).Maximum likelihood from Incomplete Data using the EM Algorithm, Dempster, Laird, and Rubin (Journal of the Royal Statistical Society, Series B, 1977) (Maximum Likelihood from Incomplete Data via the EM Algorithm). In the mid 1980s, ML took a dramatic turn, along with AI, towards the widespread use of probabilistic and statistical methods. One of the most influential models of machine learning during the 1990s was based on Fisher’s notion of maximum likelihood estimation. Since most interesting probabilistic models in AI had latent (unobserved) variables, maximum likelihood could not be directly applied. The EM algorithm, popularized by three Harvard statisticians, came the rescue. It is probably the most widely used statistical method in ML in the past 25 years, and well worth knowing. This paper, which is cited over 50,000 times on Google Scholar, requires a certain level of mathematical sophistication, but it is representative of modern ML, and much of the edifice of modern ML is based on ideas like EM. A very simple way to think of EM is in terms of “data hallucination”. Let’s say you want to compute the mean of 20 numbers, but forgot to measure the last 5 numbers. Well, you could compute the mean over the 15 numbers only, or you could do something clever, namely put in an initial guess of the mean for each of the missing 5 numbers. This leads to an easy recurrence relation that lets you find the true mean. In the one dimensional case, this happens to be the same as ignoring the last 5 numbers, but in the two dimensional case, where one or the other dimension may be different, EM finds a different solution.A Theory of the Learnable, by Les Valiant, Communications of the ACM, 1984. (https://people.mpi-inf.mpg.de/~mehlhorn/SeminarEvolvability/ValiantLearnable.pdf). George Orwell wrote a brilliant novel about the rise of the all powerful all knowing Government, which spies on everyone. Well, in the same year of the novel, Les Valiant, a brilliant computer scientist at Harvard proved that Orwell’s fears could not be completely realized due to intrinsic limitations on what can be learned from data in polynomial time. That is, even if the Government could spy on individuals, it is possible to construct functions whose identity may be hidden because it would require intractable computation to discover them. Valiant’s work lead to his winning the Turing award several decades later, computer science’s version of the Nobel Prize. What Valiant did in this landmark paper was articulate a theory of machine learning that is analogous to complexity theory for computation. He defined PAC learning, or probably approximately correct learning, as a model of knowledge acquisition from data, and showed examples where a class of functions was PAC learnable, and also speculated about non-learnable functions. Valiant’s work in the past three decades has been hugely influential. For example, the most widely used ensemble method in ML is called boosting, and came out as a direct result of PAC learning. Also to be noted is that support vector machines or SVMs were justified using the tools of PAC learning. This is a short but beautifully written paper, and while it is not an easy read, your ability to understand and grasp this paper will make the difference between whether you are a ML scientist or an ML programmer (not to make any value judgements of either, the world needs plenty of both types of people!).Intelligence without representation, Rodney Brooks, IJCAI 1987 Computers and Thought Award lecture (http://www.fc.uaem.mx/~bruno/material/brooks_87_representation.pdf). Brooks based his ideas for building “behavior-based robots” on ethology, the study of insect behavior. What ethologists found was that ants, bees, and lots of other insects were incredibly sophisticated in their behaviors, building large complex societies (ant colonies, bee hives), and yet their decision making capacity seemed to be based on fairly simple rules. Brooks took this type of idea to heart, and launched a major critique at the then representation heavy apparatus of modern knowledge-based AI. He argued that robots built using knowledge-based AI would never function well enough in the real world to survive. A robot crossing the road that sees a truck and begins to reason about what it should do would get flattened by the truck before its reasoning engine came up with a decision. According to Brooks, this failure was due to a misunderstanding of how brains are designed to produce behavior. In animals, he argued, behaviors are hard wired in a layered highly modularized form, so that complexity emerges from the interleaving of many simple behaviors. One of his early PhD students, Jonathan Connell, showed that you can design a complex robot, called Herbert (after Herb Simon), that could do a complex task of searching an indoor building for soda cans and picking them and throwing them into trash, all the while having no explicit representation anywhere of the task. Later, after Jon Connell graduated, he came to work for IBM Research, where he and I collaborated on applying RL to teach behavior-based robots new behaviors. Brooks was a true pioneer of robotics, and inserted a real-world emphasis in his work that was till then sorely lacking. He had a common-sense wisdom about how to apply the right sort of engineering design to a problem, and was not enamored of using fancy math to solve problems that had far simpler solutions. Much of the success of modern autonomous driving systems owes something to Brooks’ ideas. It is possible that the tragic accident in Arizona involving an Uber vehicle might have been averted had that particular vehicle been outfitted with a behavior-based design (which countermands bad decisions, like the one the Uber vehicle allegedly made, of labeling he pedestrian as a false positive).Natural Gradient Works Efficiently in Learning, Amari, Neural Computation, 1989 (http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.452.7280&rep=rep1&type=pdf). One of the living legends of statistics is the Indian scientist C.R. Rao, now in his 90s, who basically has done the most since Fisher in building up the edifice of modern statistics. C. R. Rao invented much of modern multivariate statistics as a young researcher at Univ. of Cambridge, England, due to his study of fossils of human bones from Ethiopia. In a classic paper written in his 20s, C. R. Rao showed that the space of probability distributions is curved, like Einstein’s space-time, and has a Riemannian inner product defined on the space of tangents at each point on its surface. He later showed how the Fisher information metric could be used to define this inner product. Amari, a brain science researcher in Japan, used this insight to define natural gradient methods, a widely used class of methods to train neural networks, where the direction pursued to modify the weights at any given point is not the Euclidean direction, but the direction that is based on analyzing the curved structure of the underlying probability manifold. Amari showed natural gradient often works better, and later wrote a highly sophisticated treatise on information geometry, expanding on his work on natural gradients. Many years later, in 2013, a group of PhD students and I showed that natural gradient methods could actually be viewed as special cases of a powerful class of dual space gradient methods called mirror descent, invented by Russian optimization researchers Nemirovksy and Yudin. Mirror descent has now become a basis for one of the most widely used gradient methods in deep learning called ADAGRAD by Duchi (now at Stanford), Hazan (now at Princeton) and Singer (now at Google). It is very important to understand these various formulations of gradient descent methods, which requires exploring some beautiful connections between geometry and statistics.Learning to Predict by the Methods of Temporal Differences, by Richard Sutton, Machine Learning journal, pp 9–44, 1988 (https://pdfs.semanticscholar.org/9c06/865e912788a6a51470724e087853d7269195.pdf). TD learning remains the most widely used reinforcement learning method, 34 years after they were invented by UMass PhD student Richard Sutton, working in collaboration with his former PhD advisor, Andrew Barto, both of whom can be said to have laid the foundations of the modern field of RL (on whose work the company Deep Mind was originally formed, and then acquired by Google). It is worth noting that Arthur Samuel in 1950s experimented with a simple form of TD learning, and used it to teach an IBM 701 to play checkers, which can be said to be the first implementation of both RL and ML in the modern era. But Rich Sutton brought TD -learning to life, and if you read the above paper, you’ll see the mathematical sophistication he brought to its study was far beyond Samuel. TD learning is now far beyond this paper, and if you want to see how mathematically sophisticated its modern variants are, I will point you to the following paper (which builds on the work of one of my former PhD students, Bo Liu, who brought the study of gradient TD methods to a new level with his work on dual space analysis). Janet Yu has written a very long (80+ pages) dense mathematical treatise on the modern version of gradient TD, which you have to be very strong in math to understand fully ([1712.09652] On Convergence of some Gradient-based Temporal-Differences Algorithms for Off-Policy Learning). TD remains one of the few ML methods for which there is some evidence that it is biologically plausible. The brain seems to encode TD error using dopamine neurotransmitters. The study of TD in the brain is a very active area of research (see http://www.gatsby.ucl.ac.uk/~dayan/papers/sdm97.pdf).Human learning in Atari, Tsividis et al., AAAI 2017 (http://gershmanlab.webfactional.com/pubs/Tsividis17.pdf). Deep reinforcement learning was popularized in a sensational paper in Nature (Human-level control through deep reinforcement learning) by a large group of Deep Mind researchers, and it is by now so well known and cited that I resisted the temptation to include it in my top 20 list (where most people would put it). It has led to large numbers of follow on papers, but many of these seem to miss the fairly obvious fact that there is a huge gulf between the speed at which humans learn Atari games and TD Q-learning with convolutional neural nets does so. This beautiful paper by cognitive scientists at MIT and Harvard shows that humans learn many of the Atari games in a matter of minutes in real time play, whereas deep RL methods require tens of millions of steps (which would be many months of human time, perhaps even years!). So, deep RL cannot be the ultimate solution to the Atari problem, even if it is currently perhaps the best we can do. There is a huge performance gap between humans and machines here, and if you are a young ML researcher, this is where I would go to make the next breakthrough. Humans seem to do much more than deep RL when learning to play Atari.Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers, Boyd et al., Foundations and Trends in Machine Learning, 2011 (Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers, which has MATLAB code as well). The 21st century has arrived, and with it, the dawn of cloud computing, and machine learning is poised to exploit these large numbers of cloud based computational structures. This very long and beautifully written paper by Stanford optimization guru Stephen Boyd and colleagues shows how to design cloud based ML algorithms using a broad and powerful framework called Alternating Direction Method of Multipliers (or ADMMs). As the saying goes in the Wizard of Oz, “we are no longer in Kansas, Toto”. Namely, with this paper, we are now squarely in modern machine learning land, where the going gets tough (but, then as the saying goes, “the tough get going”). This is a mathematically deep and intense paper, of more than 100 pages, so it is not an easy read (unless, that is, your are someone like Walter Pitts!). But, the several weeks or months you spend reading it will greatly improve your ability to see how to exploit modern optimization knowledge to speed up many machine learning methods. What is provided here is a generic tool box, and you can design many specialized variants (including Hadoop based variants, as shown in the paper). To understand this paper, you need to understand duality theory, and Boyd himself has written a nice book on convex optimization to help you bridge that chasm. The paper is highly cited, for good reason, as it is a model of clarity.Learning Deep Architectures for AI, by Bengio, TR 1312, Univ. of Montreal (https://www.iro.umontreal.ca/~lisa/pointeurs/TR1312.pdf) (also a paper published in the journal Foundations and Trends in Machine Learning). Bengio has done more than almost anyone else in popularizing deep learning, and is also one of its primary originators and innovators. In this paper, he lays out a compelling vision for why AI and ML systems need to incorporate ideas from deep learning, and while many of the specifics he says have changed due to the rapid progress in deep learning in the last few years, this paper is a classic that bears well. This paper was written as counter point to the then popular approach of shallow architectures in machine learning, such as kernel methods. Bengio is giving another of his popular tutorials on deep learning at the forthcoming IJCAI conference in July in Sweden, in case you are interested in attending the conference or the tutorial. I don't have to say much more about deep learning, as it is the subject of a barrage of publicity these days. Suffice it to say that today AI is very much in the paradigm of deep learning (meaning a framework in which every problem is posed as a problem of deep learning, whether it is the right approach or not!). Time will tell how well deep learning survives in its current form. There are beginning to be worries about the robustness of deep learning solutions (the Imagenet architectures seem very vulnerable to random noise, which humans can’t even see, let along respond to), and the sample complexity seems formidable still. Scalability remains an open question, but deep learning has shown remarkable performance in many areas, including computer vision (if you download the latest version of MATLAB R2018a, you can run the demo image recognition program with a web cam with objects in your own house, and decide for yourself how well you think deep learning works in the real world).Theoretical Impediments to Machine Learning, with Seven Sparks from the Causal Revolution, by Judea Pearl, Arxiv 2018. ([1801.04016] Theoretical Impediments to Machine Learning With Seven Sparks from the Causal Revolution). Pearl is in my view the Isaac Newton of AI. He developed the broad probabilistic framework of graphical models, which dominated AI in the 1990s-2010s. He subsequently went into a different direction with his work on causal models, and now argues that probabilities are “an epiphenomenon” (or a surface property, of a much deeper causal truth). Pearl’s work on causal models has yet to gain the same traction in AI as his earlier work on graphical models (which is a major subfield in both AI and ML). Largely, the reasons have to do with the sort of applications that causal models fit well with. Pearl is focusing on domains like healthcare, education, climate change, societal models etc. where interventions are needed to change the status quo. In these hugely important practical applications, he argues that descriptive statistics is not the end goal, but causal models are. His 2009 2nd edition of Causality is still the most definitive modern treatment of the topic, and well worth acquiring.Prospect Theory: An Analysis of Decisions under Risk, by Daniel Kahneman and Amos Tversky, Econometrica, pp. 263–291, 1979. Daniel Kahneman received the Nobel prize in Economics for this work, with his collaborator Amos Tversky (who sadly died, and could not share in the prize). In this pathbreaking work, they asked themselves the simple question: how do humans make decisions under uncertainty? Do they follow the standard economic model of maximizing expected utility? If I gave you the choice between two outcomes: choose Door 1, and with 50% probability, you get no cash prize, or you get $300; alternatively if you choose Door 2, you get a guaranteed prize of $100. It perhaps won't surprise you that many humans choose Door 2, even through expected utility theory shows you should choose Door 1( since the expected utility is $150, much higher than Door 2). What’s going on? Well, humans tend to be risk averse. We would rather have the $100 for sure, than risk getting nothing with Door 1. This beautiful paper, which has been cited over 50,000 times, explores such questions in a number of beautiful simple experiments that been repeated all over the world with similar results. Well, here’s the rub. Much of the theory of modern probabilistic decision making and reinforcement learning in AI is based on maximizing expected utility (Markov decision processes, Q-learning, etc.). If KT is right, then much of modern AI is barking up the wrong tree! If you care about how humans actually make decisions, should you continue to chose an incorrect approach? Your choice. Read this paper and decide.Towards an Architecture for Never-ending Language Learning, Carlson et al., AAAI 2010. Humans learn over a period of decades, but most machine learning systems learn over a much shorter period of time, often just a single task. This CMU effort led by my former PhD advisor, Thomas Mitchell, explores how a machine learning system can learn over a very long period of time, by exploring the web, and learning millions of useful facts. You can interact with the actual NELL system online at Carnegie Mellon University. NELL is a fascinating example of how the tools of modern computer technology, namely the world wide web, makes it possible to design ML systems that can run forever. NELL could potentially live longer than any of us, and constantly acquire facts. One issue, at the heart of recent controversies, is “fake news”, of course. How does NELL know what it has learned is true? The web is full of fake assertions. NELL currently uses a human vetting approach of deciding which facts it learns are really to be trusted. Similar systems can be designed for image labeling, language interactions, and many others.Topology and Data, by Gunnar Carlson, Bulletin of the American Mathematical Society, April 2009 (http://www.ams.org/images/carlsson-notes.pdf). The question that many researchers are interested in knowing the answer to is: where is ML going in the next decade? This well known Stanford mathematician is arguing in favor of the use of more sophisticated methods from topology, a well developed area of math that studies the abstract properties of shape. Topology is what mathematicians use to decide that a coffee cup (with a handle) and a doughnut are essentially the same, since one can be smoothly deformed into the other without cutting. Topology has one great strength: it can be used to analyze data even when standard smoothness assumptions in ML are not possible to make. It goes without saying that the mathematical sophistication needed here is quite high, but Carlson refrains from getting very deep into the technical subject matter, giving for the most part, high level examples of what structure can be inferred using the tools of computational topology.2001: A Space Odyssey, book by Arthur C. Clarke, and movie by Stanley Kubrick. My next and last choice of reading — this has gone on long enough, and both you and I are getting a bit tired by now — is not an AI paper, but a movie and the associated book. The computer HAL in Kubrick’s movie 2001 is to my mind the best exemplar of an AI based intelligent system, one that is hopefully realizable soon. 2001 was released in 1968, exactly 50 years from now, and its 50th anniversary was marked recently. Many of my students and colleagues, I find, have not seen 2001. That is indeed a sacrilege. If you are all all interested in AI or ML, you owe it to yourself to see this movie, or read the book, and preferably do both. It is in my mind the most intelligent science fiction movie ever made, and it puts all later movies to shame (no, there is no silly laser sword fights or fake explosions or Darth Vaders here!). Instead, Stanley Kubrick designed the movie to be as realistic as the technology from 1960s would allow, and it is surprisingly modern even today. HAL is of course legendary from his voice (“I’m sorry Dave” is now available as a ring tone on many cellphones). But, HAL is also a great example of how modern AI will work with humans, and help assist many functions. Many long voyages into space, such as Mars or beyond, cannot be done with a HAL, as humans will have to sleep or be in hibernation to save on storage for food etc. There is a nice book by Stork on doing a scene by scene analysis of HAL in the movie, with where AI is in the 21st century. This book is also worth acquiring.OK, I’m ending this uber long reply two papers short of the required 20, but I’m sure I’ve given you plenty of material to read and digest. I did also cheat a bit here and there, and gave you multiple papers to read per bullet. Happy reading. Hope your journey into the fascinating world of AI is every bit as rewarding and fun as mine has been over the past 30+ years.
- Home >
- Catalog >
- Miscellaneous >
- Cover Letter Examples >
- Simple Cover Letter Examples >
- Basic Cover Letter >
- This Can Be Used As A Simple Cover Letter Insert