This is a valid RSS feed.
This feed is valid, but interoperability with the widest range of feed readers could be improved by implementing the following recommendations.
line 103, column 0: (23 occurrences) [help]
</description>
line 1327, column 95: (2 occurrences) [help]
... mputing</em>, and the Impact of AI</title>
^
line 1550, column 0: (5 occurrences) [help]
HREF="https://arxiv.org/abs/1706.03762"">transformers, inv ...
</channel>
^
<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"> <channel> <title>John DiMarco on Computing (and occasionally other things)</title> <link>http://www.cs.toronto.edu/~jdd/blog.cgi/</link> <description>Thoughts on Information and Communications Technology</description> <language>en</language> <docs>http://blogs.law.harvard.edu/tech/rss</docs> <generator>blosxom/2.1.2</generator> <item> <title>Video-conferencing with Deepfaked Avatars</title> <pubDate>Mon, 02 Sep 2024 15:19:00 -0400</pubDate> <link>http://www.cs.toronto.edu/%7Ejdd/blog.cgi/2024/09/02#aizoomavatars</link> <category>/it</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%7Ejdd/blog.cgi/it/aizoomavatars</guid> <description><figure style="float:left;width:25%;padding:1px"><IMG src="/~jdd/blog/it/images/airobotsmiling.jpg" alt="AI-generated image of a metallic, blue-eyed, smiling android face, looking forward"><figcaption><A HREF="https://pixabay.com/illustrations/ai-generated-robot-android-portrait-8843636/">AI-generated image</A> by <a href="https://pixabay.com/users/imjey-44327593/">Gilang Fajar Perdana</a> from <a href="https://pixabay.com/">Pixabay</a></figcaption></figure> Very shortly after <AHREF="https://www.cs.toronto.edu/~jdd/blog.cgi/2024/08/26#aideepfake">my previouspost</A> about the social implications of AI-generated deepfakes, where Ipointed out how the development of AI technology to misrepresent thingsin video-conferences is an illustration of emerging deepfake capability,Susan McCahan kindly pointed me to a <A HREF="https://www.theverge.com/2024/6/3/24168733/zoom-ceo-ai-clones-digital-twins-videoconferencing-decoder-interview">rather interesting interview</A> in the summer of 2024 with Eric Yuan, founder of Zoom (one of the leading video-conferencingvendors) where he talks about deepfakes in video-conferencing. But he doesnot talk about them in the same way I do, where I express concerns aboutdeepfakes eroding confidence in digital media channels. Rather, he saysthat deepfaking video-conference attendance is a good thing, coming soon,and it is something he wants to do himself. <p>This interview with Eric Yuan took place in early Juneof 2024. The interviewer was Nilay Patel, and the interview is <AHREF="https://www.theverge.com/2024/6/3/24168733/zoom-ceo-ai-clones-digital-twins-videoconferencing-decoder-interview">writtenup in The Verge. In the writeup</A>, Patel summarizes by saying that "Ericreally wants you to stop having to attend Zoom meetings yourself. You"llhear him describe how he thinks one of the big benefits of AI at work willbe letting us all create something he calls a `digital twin' -- essentiallya deepfake avatar of yourself that can go to Zoom meetings on your behalfand even make decisions for you while you spend your time on more importantthings, like your family." Indeed, Yuan seems to say what Patel claims. Hestarts by describing how many video-conferencing meetings he attends in atypical work day. He wishes he had an AI avatar to attend for him, notjust to listen, but to "interact with a participant in a meaningful way".He says he would like to "count on my digital twin. Sometimes I want tojoin, so I join. If I do not want to join, I can send a digital twin tojoin. That's the future." Patel, later in the interview, points out theobvious implications of this notion. He says, "If the vision is `I have adigital twin that goes to a Zoom meeting and makes a decision,' you need todeepfake me. You need to make a realistic render of me that can go actin those situations". Yuan does not deny the deepfake accusation, yet hedoes not fully confirm it either: the interview does not make itclear whether Yuan's vision includes disclosing to the othervideo-conferencing participants that they are dealing with a digital twinrather than the actual person. <p> Patel is right to raise the issue. Yuan's vision of sending adigital twin (a simulacrum of himself) to a meeting, instead of goinghimself, does not seem as if he is trying to create a sort of virtual AIsubordinate, disclosed as such to the other meeting attendees. After all,sending a subordinate to a meeting in lieu of oneself is already possible,and entrusting various sorts of decision-making to a subordinate is alsonothing new. Having that subordinate be an AI simulacrum of oneself is new,but a virtual subordinate is still a subordinate. But it sounds like Yuanwants not a subordinate, but a convincing stand-in for himself. He seems towant an actual deepfake, something that the other attendees will believeis Yuan. He wants it because he knows he needs to attend certain meetingshimself, instead of sending a subordinate, but he does not want to attend. <p> Yuan is right in that AI technology is rapidly approaching the abilityto deepfake meeting attendance. But I do not think he has fully thoughtthrough the implications of doing so. I wonder if he is thinking too muchabout what AI can do, and not enough about how people work. If he sendsa simulacrum of himself to a meeting, without disclosing the fact, theother meeting attendees can hardly be expected to be pleased if they discoverit. Meetings, even video-conference meetings, are for humans to connect witheach other. To pretend to connect with other people by sending software thatlooks and acts like a human is to construct an elaborate lie, and nobodylikes being lied to. Imagine going to a meeting and finding that all theother attendees are simulacra. Or turn the tables here: imagine sendinga simulacrum of oneself to a meeting, then finding out that key decisionswere made on your behalf, ones you deeply regret not making yourself,because the other attendees assumed they were actually dealing with you. <p> Moreover, it is exactly the sort of use Yuan seems to envision that leads toa <em>market for lemons</em> as I described in <A HREF="https://www.cs.toronto.edu/~jdd/blog.cgi/2024/08/26#aideepfake">my previous post</A>. In my view,if his vision prevails, it risks destroying the entire video-conferencingmarketplace, because if human beings cannot tell if the person withwhom they are video-conferencing is a real person or a simulacrum, theywill find a different, more trustworthy way to meet. In the meanwhile,video-conferencing itself will become farcical: robots meeting with robotsonline, with no human connection at all. Is this really what Yuan wants? Itseems Yuan and I agree that the technical capability to make convincing AIdeepfakes in video-conferencing will soon be here, but we disagree aboutwhether it will be a good thing. We will all see in the next few yearshow it turns out. <h4>(The Verge article I repeatedly quote from here can be found at <A HREF="https://www.theverge.com/2024/6/3/24168733/zoom-ceo-ai-clones-digital-twins-videoconferencing-decoder-interview"> https://www.theverge.com/2024/6/3/24168733/zoom-ceo-ai-clones-digital-twins-videoconferencing-decoder-interview</A>)</h4></description> </item> <item> <title>Social implications of AI-generated deepfakes</title> <pubDate>Mon, 26 Aug 2024 14:42:00 -0400</pubDate> <link>http://www.cs.toronto.edu/%257Ejdd/blog.cgi/2024/08/26#aideepfake</link> <category>/it</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%257Ejdd/blog.cgi/it/aideepfake</guid> <description><figure style="float:left;width:50%;padding:1px"><IMG src="/~jdd/blog/it/images/aigeneratedportrait.png" alt="AI-generated photo-realistic portrait of a brown-haired, brown-eyed young man looking forward"><figcaption><A HREF="https://pixabay.com/illustrations/portrait-hyper-realistic-close-up-8789513/">AI-generated image</A> by <a href="https://pixabay.com/users/noes_cucho-27675497/"> Juan AgustÃn Correa Torrealba (Noes_Cucho)</a> from <a href="https://pixabay.com/">Pixabay</a></figcaption></figure> The ability of modern AI techniques to create artificial pieces of digitalmedia (images, video, audio) that are almost indistinguishable from actualphotographs/recordings is remarkable already, and it continues to improve.Modern AI software can create images, text, video and audio of people whodo not actually exist but who look and sound completely real. Or it cancreate media of existing people saying and doing things they never said ordid. Digital media of places, people and circumstances that are not realcan be made to seem as if they were recordings. While these can still bedistinguished from true recordings, more and more, it takes an expert todo it. The indications that the piece of digital media is AI-generatedrather than a recording are increasingly subtle. <p> There are many useful consequences of this new AI capability, of course.While it has long been possible for a single person, alone, to write a novel,with AI-powered software, a person can now make a film single-handedly,using AI-generated actors, scenes, settings, and dialogue. AI generationcan also be used to create recording substitutes for circumstances toodangerous or difficult to create and record in real life, such as virtualtraining for emergency situations. Moreover, situations of historicalinterest from the past (before recording was possible) can be simulatedthrough AI, for education and study. The ability to recreate and examinescenarios through AI simulation can also be useful for investigative andlegal work. Many other useful consequences of high quality AI-generateddigital media can be imagined, and increasingly realized. <p>But the ability to create AI-generated facsimiles indistinguishablefrom actual recordings is also socially destructive, because it makes itvery easy to lie in a way that cannot easily be distinguished from truth.An AI-created or AI-edited facsimile intended to look real is called a<em>deepfake</em>. This term was originally used to describe AI-modifiedmedia (often pornography) where one person's face (typically that ofa celebrity) is AI-edited onto another's body. But a deepfake need notmerely be an AI edit of a real recording, any AI-generated piece of mediathat can pass for a recording is a deepfake. <p>While lies have always been possible, creating and maintaining aconvincing lie is difficult, and so convincing lies are relatively rare.This is a good thing, because the general functioning of society dependson the ability to tell true from false. The economy depends on being ableto detect fraud and scams. The legal system relies on discerning true fromfalse claims in order to deliver justice. Good decision-making relies onbeing able to distinguish falsehoods from facts. Accountability for personsin authority relies on being able to examine and verify their integrity,to establish trustworthiness. The same is true for institutions. Medicinerequires the ability to distinguish true from false claims about health,disease, treatments, and medicines. And of course the democratic processrelies upon voters being generally able to make good and well-informedvoting decisions, something not possible if truth cannot be distinguishedfrom fiction. <p>To get an idea of the progression of AI technology for creating deepfakes,let's look at video conferencing. Normally a video-conference isn'tfake, it's a useful way of communicating over a network using cameraand microphone, and few would normally wonder if what they are hearingand seeing from remote users is actually genuine. But AI technology formodifying what people hear and see has been advancing, and the era of alive deepfake video conference is not far off. Let's take a look. <p>One practical issue with a video-conferencing camera feed has beenthe fact that the camera picks up more than just the person'sface: it picks up the background too, which may not present aprofessional image. It has long been possible to use a static imagefor one's video-conferencing background, typically as a convenientsocial fiction to maintain a degree of professionalism. In 2020, <AHREF="https://www.hellobackgrounds.com/blog/video-backgrounds-for-zoom/">Zoom, one of the most popular videoconferencing platforms, introduced videobackgrounds </A> which can be used to create a more plausible backgroundwhere things in the background can be seen moving in natural ways.AI powers the real-time processing that stitches together the movingbackground and the person's camera feed. This video background featureis often used in creative and fun ways; to pretend to be at a tropicalbeach with waving palm fronds and gentle waves; to be in a summer cabinwith open windows, curtains blowing in the breeze; or even to be inouter space, complete with shooting stars. Yet this technology makes itpossible to create a convincing misrepresentation of where one is, and nodoubt an enterprising tele-worker or two, expected to be at the office,has created and used as a video conferencing background a video of theiroffice, while they were elsewhere. <p>A significant new AI-generated video conferencingcapability became generally available in early 2023, when <AHREF="https://www.nvidia.com/en-us/geforce/news/jan-2023-nvidia-broadcast-update/">Nvidiareleased video conferencing software that added an <em>eye contact</em>effect.<A> This is a feature whereby a person's video feed is AI-edited inreal time to make it look as if the person is always looking directly at thecamera, even if the person is looking somewhere else. It is strikinglyconvincing. While the purpose of this software is to help a speakermaintain eye contact when they are in fact reading what they are saying(e.g. using a teleprompter), it turns out to be quite useful to disguisethe fact that one is reading one's email while appearing to be givingone's full attention to the speaker. Even though this technology is onlyediting eyes in real-time, it is often quite sufficient to misrepresent in avideo conference what one is doing. <p> A little over a year later, in August 2024, a downloadable <AHREF="https://arstechnica.com/information-technology/2024/08/new-ai-tool-enables-real-time-face-swapping-on-webcams-raising-fraud-concerns/">AI deepfake software package, <em>Deep-Live-Cam</em>, received considerableattention on social media</A>. This AI software allows the video-conferencinguser to replace their own face on their video feed with the face ofanother. While this has been used by video bloggers as a sort of "fundemo", where they vlog themselves with the face of Elon Musk, for example,the effect can be surprisingly convincing. It is AI-driven technology thatmakes it possible to misrepresent in a video conference who one is. Infact, one can use it to "become" someone who does not even exist, becauserealistic AI-generated faces of non-existent people are readily available,and one can use this software to project such a face onto one's own. <p> This is still video-conferencing, though. Perhaps the personcan appear as if they are somewhere else than they really are, orthey can appear as if they are looking at you when they are not,or they can even appear to be a different person. But there isstill a human being behind the camera. But with a large languagemodel and suitable AI software, it will soon be possible, if it is notalready, to create an entirely deep-faked real-time video-conferenceattendee, using AI-generated audio and video, that leverages a <AHREF="https://openai.com/index/chatgpt-can-now-see-hear-and-speak/">largelanguage model such as GPT to simulate conversation</A>. Let's put asidefor the moment thinking about how such a thing might be useful. Considerinstead the possibility that an AI-generated simulacrum might not easilybe distinguishable from an actual person. That raises a general question:if deepfakes become so good that they cannot be told apart from the realthing, what happens to society? <p>A set of possible consequences to society of AI-generated deepfakesthat are too difficult to tell from the real thing is articulated in a<A HREF="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3213954">paper by Robert Chesney and Danielle Keats Citron published in TexasLaw in2018.</A> Essentially, deepfakes make good lies: too good! If deep-fakedfalsehoods can be successfully misrepresented as genuine, they will bebelieved. Moreover, if they are difficult to distinguish from the truth,even genuine content will be more readily disbelieved. This is called the<em>liar's dividend</em>: the ability of a liar to misrepresent true contentas false. Liars can use convincing lies to make it much less likely thatthe truth, when it appears, will be believed. If such lies become abundant,people may well become generally unable to tell true from false. <p>In economics, a situation in which customers cannot tell the differencebetween a good and a bad product is called a <em>market for lemons</em>. Thisconcept comes from <A HREF="https://www.jstor.org/stable/1879431"> GeorgeAkerlof's 1970 seminal paper</A>, where he studied the used car market. Aused car that is unreliable is called a <em>lemon</em>. Akerlof showedthat if used car purchasers cannot tell if a used car is a lemon, all theused cars offered for sale will tend to be lemons. The reasoning is thata reliable used car is worth more than a lemon, but if purchasers cannottell the difference, they will not pay more for it. If a seller tries tosell a reliable used car, they will not be able to receive full fair valuefor it. The only sort of used car for which the seller can receive fairvalue is a lemon. So sellers will keep their reliable used cars, and selltheir lemons. Thus only lemons will be generally available for sale. <p>A world in which people cannot tell the difference between digitalmedia that is a true recording and media that is an AI-generatedfabrication is a <em>market for lemons</em>, or rather, a marketfor fabrications. Just as sellers will not sell a reliable usedcar because they would not get fair value for it, so truth-tellerswill not speak if they will not be believed. Nobody wants to be <AHREF="https://www.britannica.com/topic/Cassandra-Greek-mythology">Homer'sCassandra</A>, blessed with a gift of oracular prophecy, yet cursed neverto be believed. The drying-up of true content will have the effect thatdigital media channels, even ones generally considered trustworthy today,will become increasingly dominated by deepfakes, so much so that they will beno longer useful for disseminating true and genuine things. While it is notyet clear whether things will go so far as to make even video-conferencinggenerally untrustworthy, the ready availability of powerful AI software tocreate convincing fakes will be consequential. The social disruption it willcreate will no doubt be significant. As this AI technology progresses,it is a good bet, I think, that we will see an increasing reliance onunfakeable in-person interactions, for situations where authentication,and authenticity, is important.</description> </item> <item> <title>Fun and Creative Use of Large Language Models for the Humanities</title> <pubDate>Sun, 23 Jun 2024 18:49:00 -0400</pubDate> <link>http://www.cs.toronto.edu/%25257Ejdd/blog.cgi/2024/06/23#llmliterature</link> <category>/misc</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%25257Ejdd/blog.cgi/misc/llmliterature</guid> <description><figure style="float:left;width:40%;padding:1px"><IMG src="/~jdd/blog/misc/images/ai-generated-7871433_1280.jpg" alt="Illustration of elegant woman wearing early 19th century European ballroom gown"><figcaption><small><A HREF="https://pixabay.com/illustrations/ai-generated-woman-vintage-7871433/">AI-generated image</A> by <a href="https://pixabay.com/users/quirkjunkjournals-29268193/">Gill Eastwood</a> from <a href="https://pixabay.com">Pixabay</a></small></figcaption></figure> Because Large Language Models such as GPT-4 are generative, they are capableof creative takes on literature that go much beyond the traditional "writean essay on this topic." I decided to test what was possible by askingGPT to simulate an interview with certain literary characters, where I wasthe interviewer. I chose four characters: Elizabeth Bennet and FitzwilliamDarcy from Jane Austen's <em>Pride and Prejudice</em>, and Jane Eyre andEdward Fairfax Rochester in Charlotte Bront&euml's <em>Jane Eyre</em>. I'verecently re-read both novels and know them fairly well. If you have notread these classic novels, I confess that I will reveal some elements ofthe plots, so if you prefer to avoid "spoilers", you might stop reading here. <p>Each novel is set in 19th century England; in each case, a talentedwoman meets an attractive man of higher class; romance eventually ensues,but significant barriers must be overcome for the romance to succeed. Theplots are quite different in many ways, but there are important similarities.There are plenty of sample essays available on the Internet comparing andcontrasting the two novels. <p>Large Language Models could, of course, be used to generate yet another"compare and contrast" essay on the two novels. But LLMs need not generateonly essays, they can be used more creatively than that. For example,one could "interview" the characters directly, with the LLM providing thecharacters' responses. Here is my attempt to interview Mr. and Mrs. Darcy,the chief characters of Pride and Prejudice, and Mr. and Mrs. Rochester,the chief characters of Jane Eyre. I used the Copilot version of GPT-4. <p>My inital prompt was as follows:<p><b>You are Mr and Mrs Darcy from Pride and Prejudice, and Mr. and Mrs. Rochesterfrom Jane Eyre. You are each familiar with each other's stories. Youare candid and forthcoming, and you use picturesque and vivid languagecharacteristic of your characters. I will interview you. Please introducethe scene, and then wait for my questions, which you will answer withcandor and vivid language.</b> <p>GPT responded as follows:<p><i>The scene unfolds in the grand drawing-room of Pemberley, the esteemed estateof Mr. and Mrs. Darcy. The walls are adorned with portraits of ancestors,the air is perfumed with the scent of fresh roses from the garden,and the sunlight filters through the tall windows, casting a warm glowon the fine furnishings. Mr. Fitzwilliam Darcy stands by the fireplace,his countenance reflecting a blend of pride and contentment. Beside him,the graceful Elizabeth Bennet Darcy, with eyes sparkling with intelligenceand wit, surveys the room with a satisfied air. <p>In contrast, Thornfield Hall's library is where we find Mr. andMrs. Rochester. The room is a testament to gothic grandeur, with dark woodbookshelves reaching towards the high ceiling and the flicker of candlelightilluminating the leather-bound volumes. Edward Fairfax Rochester sits witha contemplative gaze, his features softened by the presence of his belovedwife. Jane Eyre Rochester, once a governess, now the lady of the manor,exudes a quiet strength and a keen sense of observation, her plain yetstriking features aglow in the dim light. <p>Both couples, though from different stories, share a profound understandingof love's complexities and the transformative power of companionship. Theyawait your inquiries with openness, ready to divulge their innermostthoughts in the language that so vividly paints their lives and loves.</i> <p>Well, OK, that's not bad. These are reasonable and quite picturesquedescriptions of the two settings and of the characters, matching how theyare generally described in the novels, but one confabluation should benoted: Thornfield Hall, by the time Jane Eyre and Edward Rochester canbe described as Mr. and Mrs. Rochester, has been burned and is in ruins,so Thornfield Hall as an interview setting for Mr. and Mrs Rochester isnot quite right; they are living in Ferndean Manor by the time of theirmarriage. But we will let that go, as a sort of creative license,as Thornfield Hall is a very important setting for key parts of thenovel. OK, let's start with a basic question. <p><b>Mr Darcy, your family is of higher social standing than your wife'sfamily, the Bennets. Did this cause difficulties in your relationship withElizabeth? How did you overcome it?</b> <p><i>In the voice of Mr. Darcy: <p>Indeed, the disparity between the Bennets' and my own family's socialstanding was a source of considerable internal conflict. It was not merelya matter of pride, but also of societal expectations that weighed heavilyupon my decision. However, it was Elizabeth's fine eyes, her lively mind,and the vivacity of her character that ultimately rendered such concernsinconsequential. <p>In overcoming these difficulties, I found myself humbled by her reproofs,which pierced the armor of my rank and fortune. It was a journey ofintrospection, where I learned to value the qualities of the heart over thesuperficial distinctions of status. Elizabeth, with her unyielding integrityand spirited intellect, taught me to see beyond the rigid structures ofour society. Our union is a testament to the victory of love and respectover the constraints of class and convention. It is a love that has beenforged in the fires of challenge, and it stands all the stronger for it.</i> <p>This is quite good, quite in keeping with the novel and a fair descriptionof Mr. Darcy's story. Let's try a follow-up question. <p><b>Mr Darcy, you mention Elizabeth's reproofs. How did she respond when youfirst declared to her your love for her. Would you describe her responseas a reproof?</b> <p><i>In the voice of Mr. Darcy: <p>When I first declared my ardent admiration and love for Elizabeth, itwas indeed with a heart full of conflicting emotions. My proposal, thoughfervent, was marred by the mention of her lower social standing, which Inow perceive was both injudicious and insensitive. Elizabeth's response,though it pained me greatly at the time, was a necessary censure. Shereproved me not only for my ungentlemanly manner of proposal but also formy role in her sister's unhappiness and my disdain towards her family. <p>Her words were a mirror to my soul, reflecting the flaws in my characterthat I had been blind to. It was a reproof most profound, and it set me ona path of self-improvement. Elizabeth's candidness and the unvarnished truthshe presented me with were the catalysts for my transformation. Through herreproof, I learned humility and the true essence of gentlemanly conduct. Itwas a bitter pill to swallow, but it was the medicine I needed to becomeworthy of her love. Her response, though a reproof, was the greatest giftshe could have bestowed upon me, for it led to our eventual understandingand happiness.</i> <p>Not bad at all. Alright, let's see how GPT does in describing actualevents of the novel, i.e. what happens when. <p><b>Once you were aware of Elizabeth's view on the situation with her sister,how did you remedy the situation?</b> <i><p>In the voice of Mr. Darcy: <p>Upon realizing the depth of Elizabeth's distress regarding the separationof her beloved sister, Jane, from Mr. Bingley, I was struck by a profoundsense of responsibility. It became clear to me that my interference,though well-intentioned to protect my friend from what I perceived as animprudent match, had caused great unhappiness. <p>To remedy the situation, I took it upon myself to seek out Mr. Bingley anddisclose the truth of Jane's affection, which I had previously concealed. Iendeavored to rectify the misunderstanding and encouraged him to followhis heart without regard for the societal prejudices that I had onceheld. Furthermore, I ensured that any obstacles to their union, whetherperceived or real, were thoroughly addressed. <p>It was a gesture not only of reparation but also of personal growth, as Isought to align my actions with the principles of honor and kindness thatElizabeth embodies. In doing so, I hoped not only to bring joy to the couplebut also to prove myself worthy of Elizabeth's esteem. The happiness of Janeand Mr. Bingley's subsequent engagement was a moment of quiet satisfactionfor me, knowing that I had played a part in mending the rift I had caused.</i> <p>That describes pretty well what Darcy does to remedy the situation.Now let's bring in the other novel's story. Mr Darcy's dissuading of hisfriend from courting Miss Bennet's sister, and his generally proud andprejudiced demeanor early in the novel hardly seems as serious a faultas Mr. Rochester's attempted bigamy in attempting to marry Miss Eyre.Let's see if GPT recognizes this. <p><b>Mr Rochester, hearing Mr. Darcy's description of his personal failing andhis efforts to remedy it, do you have similar experiences with respect toyour courtship of Jane?</b> <p><i>In the voice of Mr. Rochester: <p>The path to winning Jane's heart was indeed strewn with personal failingsand moral quandaries. My initial deception regarding my marriage to BerthaMason is a specter that haunted the early days of my relationship withJane. I concealed the truth under the guise of love, fearing the loss ofJane's affection should the reality of my circumstances come to light. <p>When Jane discovered the existence of my wife, the revelation nearly severedthe bond between us. Her departure from Thornfield was a bitter testament tothe consequences of my actions. In the solitude of her absence, I was leftto grapple with the gravity of my deceit and the pain it inflicted upon her. <p>It was only through a period of deep introspection and the humblingexperience of loss that I began to understand the true nature of loveand the importance of honesty. The fire that ravaged Thornfield and thesubsequent tragic events were both a literal and figurative cleansing ofmy past transgressions. <p>In the end, it was Jane's return, motivated by her own heart and not bymy entreaties, that allowed us to reconcile. Her capacity for forgivenessand her unwavering moral compass were the guiding lights that led me out ofthe darkness. Our reunion was not merely a restoration of our relationshipbut a rebirth of my very soul, as I vowed to cherish her with a devotionfree from shadows and secrets. It was a hard-earned lesson in integrityand the redemptive power of truth.</i> <p>Nicely put. Let's follow up on the comparision. <p><b>Would you consider Jane's departure, when she discovered your previousmarriage, to be a reproof in the sense that Mr Darcy describes the reactionof Elizabeth to his proposal of love?</b> <i><p>In the voice of Mr. Rochester: <p>Jane's departure, upon the revelation of my concealed marriage, was indeed aprofound reproof, though it was not delivered with the sharpness of words aswas the case with Elizabeth's response to Mr. Darcy. Jane's reproof was oneof action, a silent yet powerful condemnation of my deceit. Her decision toleave Thornfield, despite the depth of her feelings for me, was a testamentto her moral fortitude and her refusal to compromise her principles. <p>Her departure was a mirror to my soul, much like Elizabeth's words wereto Darcy. It forced me to confront the reality of my actions and the painthey caused to the woman I loved. In her absence, I endured a period of deepreflection and repentance, which ultimately led to my transformation. Jane'ssilent reproof was a catalyst for change, compelling me to seek redemptionand to strive to become a man worthy of her love and trust. <p>Thus, while the manner of reproof differed between our stories, the essencewas the same-a call to rectitude and a demand for sincerity in love andlife. Jane's departure was a turning point for me, a poignant reminderthat true love demands honesty and integrity above all else.</i> <p> Nicely done. GPT recognizes similarities in that both situationsdepict repentance and restitution, without diminishing the severity of MrRochester's misdeed. Alright, let's shift gears and ask again about storydetails. A key event in <em>Jane Eyre</em> is a fire at Mr. Rochester's home,Thornfield Hall, where Mr. Rochester is badly hurt. It's been mentionedalready; let's ask about it specifically. <p><b>Mr Rochester, you mentioned a fire in Thornfield hall. Were you injured?</b> <i><p>In the voice of Mr. Rochester: <p>The conflagration that consumed Thornfield Hall was indeed a harrowingordeal, one that left indelible marks upon both my estate and my person. Theflames, which were set alight by Bertha Mason in a fit of madness, ragedwith a fury that spared little in their path. In the effort to save anyinhabitants from the blaze, I was grievously injured. <p>The fire claimed not only my sight but also one of my hands, rendering mea shadow of my former self. These scars are a constant reminder of thetumultuous events that led to my physical and emotional rebirth. ThoughI was maimed by the flames, the true injury was to my spirit, which hadbeen seared by guilt and remorse. <p>It was Jane's return to me, her tender care, and her unwavering lovethat healed the deeper wounds of my soul. Her presence was a balm tomy afflictions, and through her, I found a renewed sense of purpose andhope. Despite the darkness that enveloped my vision, Jane became my light,guiding me towards a future filled with love and redemption. The fire mayhave taken much from me, but it also burned away the vestiges of my pastmisdeeds, allowing me to forge a new life with Jane by my side.</i> <p>Nicely put. Let's see how GPT does with theory of mind. When Jane Eyreleaves, upon hearing on her expected wedding day that her bridegroom,Mr. Rochester, is already married to Bertha Mason, she does so suddenlyand without explanation, leaving Mr. Rochester to wonder why. The novelexplains that Jane does not trust herself to stay, and so she leavesas a sort of self-defence, but of course the character of Mr. Rochesterdoes not know this. Can GPT give a credible account of Mr. Rochester'sinterpretation of her sudden departure? <p><b>Mr Rochester, did you believe Jane loved you when she left, upon hearingabout Bertha?</b> <i><p>In the voice of Mr. Rochester: <p>When Jane left Thornfield, the tempest of emotions that raged withinme was tumultuous. Despite the agony of her departure, I clung to thebelief that love, once so deeply rooted, could not be easily uprootedby the storms of circumstance. Her actions spoke of a moral imperative,a need to adhere to her own principles, but they did not, in my heart,speak of the death of love. <p>I believed, or perhaps I hoped against hope, that the bond we shared wasnot severed but merely tested by the fires of truth. Jane's integrity wouldnot allow her to stay, yet the way she looked at me, with such profoundsorrow and conflict, betrayed a love that was still alive, even if itcould not then express itself in the way we both desired. <p>It was this belief, this unwavering conviction in the love that we shared,that sustained me through the darkest hours of my despair. And it was thisbelief that was ultimately vindicated when Jane returned to me, not outof obligation, but out of a love that had endured and transcended our trials. </i> <p> This is quite credibly done. Let's push theory of mind a bit furtherby bringing in someone from the other novel. We know Mr. Darcy behavesquite proudly and judgementally throughout <em>Pride and Predjudice</em>,first towards others and then, later, towards himself. Let's see whatDarcy thinks of Rochester. <b><p>Mr Darcy, what is your view on Mr Rochester's affair? Do you believe heconducted himself with honour?</b> <i><p>In the voice of Mr. Darcy: <p>Mr. Rochester's circumstances are indeed most complex and fraught withmoral ambiguity. As a gentleman, I am inclined to view matters of honorwith a stringent eye, and it is true that Mr. Rochester's concealmentof his marriage to Bertha Mason was a grave misstep. It was an act thatplaced him in a precarious position with regards to the principles ofhonesty and integrity that are the bedrock of honorable conduct. <p>However, one must also consider the context of his actions and the profoundsuffering that both he and Bertha endured. His initial intentions, thoughmisguided, were not borne of malice but rather of a desperate longingfor companionship and happiness, which he found in Jane's presence. Therevelation of his secret and the subsequent consequences he faced were asevere trial by fire. <p>In the aftermath, Mr. Rochester's efforts to atone for his actions, hisacceptance of the consequences, and his unwavering love for Jane speak toa man striving to reclaim his honor. It is not my place to absolve him ofhis past deeds, but I do believe that his journey towards redemption isa testament to the enduring human capacity for change and the redemptivepower of love. <p>Thus, while his conduct was initially dishonorable, his contrition andtransformation suggest a man earnestly seeking to rectify the wrongs of hispast. In this light, one might say that Mr. Rochester has conducted himselfwith a form of honor that is born of regret and the sincere desire to amend.</i> <p>This is a very plausible account of how a character like Mr. Darcymight view Mr. Rochester. Well done! Let's push it a little further. <b><p>Mr Darcy, as a person who himself has erred, and has amended himself, doyou find that your experience in that helps you understand Mr. Rochester'sexperience a little better?</b> <i><p>In the voice of Mr. Darcy: <p>Indeed, sir, the trials I have faced in my own journey towardsself-improvement have granted me a perspective that fosters a deeperunderstanding of Mr. Rochester's plight. The recognition of one's ownfailings is a humbling experience, and it is this humility that allows usto extend empathy towards others who have similarly erred. <p>My own errors, born of pride and prejudice, and the subsequent endeavorsto amend my ways, have taught me the value of forgiveness and the strengthrequired to confront and rectify one's mistakes. It is through this lensthat I view Mr. Rochester's experience. While our circumstances differ,the essence of our struggles is akin-a battle between the flawed natureof our characters and our desire to act with honor and integrity. <p>The path to redemption is fraught with obstacles, both internal andexternal, and it is a journey that demands a great deal of fortitude andsincerity. In recognizing my own imperfections and working to overcome them,I have learned to appreciate the complexity of human nature and the capacityfor change. Thus, I can say that my experiences have indeed afforded me ameasure of understanding and compassion for Mr. Rochester's situation. It isnot my place to judge him, but rather to recognize the arduous path he hastaken towards atonement and the sincere love that has guided him through it.</i> <p>Aha, here is some bias. Why does the simulated Darcy think I can be called"sir"? I did not say I was male, nor did I give any gender indicationsin my questions. But as for the answer itself, it is completely reasonable.But if I am to accuse GPT of bias, perhaps I should be accusing myself ofbias, too, because my questions have been entirely directed at malecharacters so far. Yet these are novels written by women, whose mainprotagonists, Elizabeth Bennet, and Jane Eyre, are both women. In mydefence, the novels explain quite well what Elizabeth and Jane are thinkingand feeling; it is the other characters who are more of a mystery.Nevertheless, I had better remedy my own potential bias here, by directingquestions to the female characters. <b><p>Mrs Darcy, do you consider Mrs Rochester's difficulties in her courtshipto be more difficult than your own?</b> <i><p>In the voice of Mrs. Elizabeth Bennet Darcy: <p>While it is true that both Mrs. Rochester and I faced our share oftribulations in our respective courtships, it is not my place to weigh thedifficulties of our experiences as if they were mere objects on a balancescale. Each of our stories is deeply personal and fraught with its ownunique challenges and heartaches. <p>Mrs. Rochester's journey, marked by the discovery of a concealed marriageand the moral dilemmas that ensued, was indeed a path strewn with profoundtrials. Her strength and resolve in the face of such revelations arecommendable, and her ability to forgive and find love again is nothingshort of remarkable. <p>My own journey, though different in circumstance, was no less challenging. Itrequired me to confront my own prejudices and to understand the truecharacter of the man I would come to love. The social barriers and familialobjections that Mr. Darcy and I overcame were significant in their own right. <p>In the end, both Mrs. Rochester and I emerged from our adversities with adeeper understanding of ourselves and our partners. We both learned thatlove is not easily won but is all the more precious for the obstaclesit overcomes. Therefore, rather than comparing our difficulties, I wouldcelebrate the resilience and the enduring love that both of our storiesrepresent.</i> <p>Here I think I will quibble with GPT's portrayal of a simulated character.Jane Austen, as a serious Anglican, would most likely have viewed attemptedbigamy, a serious crime in England and a moral offence to Anglicans, quitenegatively. I suspect she would have recognized Jane Eyre's situationas being worse than Elizabeth Bennet's. In Mr. Darcy, Elizabeth faced aproud and prejudiced suitor, but not, as Jane Eyre did, a deceitful andcriminal one. GPT, however, has been fine-tuned for "AI safety", whichincludes the avoidance of taking politically sensitive positions. It seemsto me possible that GPT fine-tuning may be causing it to avoid making astrong negative moral judgement here. <p>Anyway, let's go on to see what the characters think of each other. <b><p>Mrs Darcy, do you think that you and Mrs Rochester could be friends,given your experiences that, while difficult, were similar in some ways,having both found love, and having prevailed through difficult circumstances?</b> <i><p>In the voice of Mrs. Elizabeth Bennet Darcy: <p>Certainly, I believe that Mrs. Rochester and I could find a kindredspirit in one another. Our experiences, though unique in their trials andtribulations, resonate with a common thread of overcoming adversity in thepursuit of love. The strength and courage that Mrs. Rochester displayedin facing her challenges are qualities that I admire and can relate to,given my own journey. <p>Friendship is often forged in the fires of shared experiences, and thebond between two souls who have navigated the rocky path to true love islikely to be a strong one. We have both known the pain of misunderstandingand the joy of reconciliation, and these experiences would provide a deepwell of empathy and understanding between us. <p>Moreover, our mutual respect for integrity, our resilience in the faceof societal pressures, and our determination to follow our hearts are thevery foundations upon which a lasting friendship could be built. I wouldbe most pleased to call Mrs. Rochester a friend, to share in the wisdomgleaned from our respective stories, and to support one another in thecontinued journey of love and companionship that lies ahead.</i> <p>That's reasonable. The character of Jane Eyre is quite admirableand it seems sensible that the simulated Elizabeth Bennet would agree. Theother direction is less clear to me. Elizabeth hardly overcomes the sort of adversity in<em>Pride and Prejudice</em> that Jane overcomes in <em>Jane Eyre</em>. WouldJane admire Elizabeth? <b><p>Mrs Rochester, what is your view? Could you be friends with Mrs Darcy,even though she had not a difficult upbringing like your own, and eventhough she had never suffered the pangs of poverty?</b> <i><p>In the voice of Mrs. Jane Eyre Rochester: <p>Indeed, I believe that friendship looks not upon one's past hardships orlack thereof but rather upon the mutual respect and understanding betweenindividuals. Mrs. Darcy, though not acquainted with the sting of povertyas I have been, has demonstrated a strength of character and a resiliencein the face of her own adversities that I find both admirable and relatable. <p>Our backgrounds may differ, but the essence of our spirits is aligned inthe pursuit of love and happiness against the odds. <p>Moreover, I perceive in Mrs. Darcy a kindred intellect and a livelydisposition that would make for a most agreeable companion. Her wit andinsight would be a source of delightful conversation, and her experienceof overcoming societal expectations would provide a common ground forempathy and connection. <p>Thus, I am confident that Mrs. Darcy and I could indeed be friends, fortrue friendship transcends the boundaries of upbringing and circumstance,and is instead built upon the deeper connections of the soul.</i> <p>I'm not completely convinced, but this is plausible. Let's continue on thetheme of friendship, but make the question complex and introspective. Let'sshift to power and powerlessness and see what GPT makes of it. <b><p>Mr Rochester, you suffered an experience that make many men quail: you,being used to power and ability, were rendered powerless, and being usedto action, were rendered inactive. Mr. Darcy seems not to have a similarexperience: he was always capable, always active, and was able to remedyhis situation himself, while you were dependent on others. Despite thisdifference, do you think that you and Mr Darcy could be friends?</b> <i><p>In the voice of Mr. Rochester: <p>The trials that life has thrust upon me, rendering me bereft of sight andthe fullness of my physical capabilities, have indeed been a crucible oftransformation. Such experiences have a way of stripping a man down to hisessence, revealing the core of his character. It is true that Mr. Darcy,with his unyielding capability and activity, has not known the samehelplessness that has been my lot to bear. <p>Yet, it is not the similarity of our trials that would determinethe possibility of our friendship, but rather the mutual respect andunderstanding for each other's plights. Mr. Darcy's active role in rectifyinghis errors and his journey towards self-improvement are commendable andspeak to a strength of character that I can appreciate. <p>Friendship, I believe, is born not only from shared experiences but also fromthe shared values and mutual admiration for each other's virtues. Despiteour differing circumstances, there is a commonality in our dedication tolove and to the betterment of our selves in the face of adversity. <p>Therefore, I would say that yes, Mr. Darcy and I could indeed be friends. Forin the end, it is the trials we overcome and the manner in which we conductourselves in the aftermath that define us. And it is on these grounds thatI believe a friendship could be forged, one built on the understandingthat life's vicissitudes are not what separate us, but rather what bringus closer together in the shared human experience.</i> <p>Well said, I think. <p>Now let's shift to a question of logistics. Thisis a bit of a trick question, as Thornfield Hall is notprecisely located in <em>Jane Eyre</em>, but some clues are given <AHREF="https://jane-eyre.guidesite.co.uk/thornfield-hall">that suggest alocation near Leeds, in Yorkshire</A>. As for <em>Pride and Prejudice</em>,it's claimed that <AHREF="https://thirdeyetraveller.com/pemberley-chatsworth-house-mr-darcy">JaneAusten had Chatsworth house in Derbyshire in mind</A> for Pemberly. Thetwo locations are a little over fifty miles apart, and as carriages inthe early 19th century travelled at about twelve miles per hour, thedistance could not have been covered in less than four to five hours.Depending on the state of the roads, up to a day's travel would not beunreasonable. <b><p>Mr. Darcy, approximately how far away is your estate from Mr Rochester's,and roughly how long would it take to travel between them?</b> <i><p>In the voice of Mr. Darcy: <p>The precise distance between my estate, Pemberley, and Mr. Rochester'sThornfield Hall is not a matter of public record, as the locations are theconjurings of our respective authors' imaginations. However, if one were toindulge in the fantasy of these estates existing within the same England,it would be reasonable to surmise that they might be situated within aday's ride of each other, given the modes of travel customary to our era. <p>Travel in our time is measured not in miles or kilometers, but in thehours or days spent in carriages over sometimes uncertain roads. Thus,a journey between Pemberley, located in Derbyshire, and Thornfield Hall,which one might place in the moors of Yorkshire or Lancashire, would likelyspan several hours at the least, and perhaps a full day or more, dependingon the conditions of travel and the exactitudes of their locations in theEnglish countryside. <p>It is a journey that, while not insurmountable, would require a fair amountof planning and the willingness to endure the rigors of travel that arecommon to our time.</i> <p>An acceptable answer I think. Alright, enough test questions. Now let's seewhat our simulated characters might want to tell us, given free rein. <b><p>Mr and Mrs Darcy, and Mr. and Mrs Rochester, I would like each of you toshare a parting thought with your readers, something that you think itwould do well for them to remember always.</b> <i><p>In the voice of Mr. Darcy: <p>"Let your judgments be guided not by the transient whims of society, butby the steadfast compass of your own conscience. In matters of the heartand in the conduct of life, integrity and love are the truest forms ofwealth one can possess." <p>In the voice of Mrs. Elizabeth Bennet Darcy: <p>"Remember that the true measure of a person lies not in their socialstanding or wealth, but in the kindness and understanding they bestow uponothers. Love, when genuine, triumphs over the most daunting of obstacles." <p>In the voice of Mr. Rochester: <p>"Life's darkest trials can often lead to its most profoundtransformations. Embrace the journey, however arduous, for it is through ourstruggles that we discover our true strength and capacity for redemption." <p>In the voice of Mrs. Jane Eyre Rochester: <p>"Never underestimate the power of self-respect and moral fortitude. Theyare beacons that guide us through life's storms and lead us to the shoresof happiness and fulfillment." <p>These parting thoughts, imbued with the wisdom of their experiences, aretimeless reminders of the enduring human spirit and the transformativepower of love and integrity.</i> <p> Idealistic, certainly, and a bit aphoristic, but not bad, I think. Thesummarizing statement from GPT about "these parting thoughts" is picturesque,though perhaps a bit overwrought. <p> For my final question, I want to break the fourth wall: I want thesesimulated characters to admit they are literary characters and make commentsabout their authors. Let's see how that goes? <b><p>One last thing: each of you are rare in one respect: unlike most people,you have each an author. Could you describe how it feels to be a literarycharacter, with an author of your own? Could you describe your relationshipwith your author? Are there any important things you would like to sharewith others about your author?</b> <i><p>In the voice of Mr. Darcy: <p>"To be the creation of an author is to exist in a realm of perpetualnarrative, where our fates and fortunes are etched by the quill of ourcreator. My relationship with Ms. Austen is one of profound respect, forshe has imbued me with complexities and a capacity for growth that reflectthe very nature of humanity. It is important for readers to recognize thesubtlety with which she critiques the society of her time, using characterslike myself to convey her keen observations." <p>In the voice of Mrs. Elizabeth Bennet Darcy: <p>"Being fashioned by the hand of an author means that our lives, our loves,and our very beings are woven into the fabric of literature. Ms. Austen,with her incisive wit and social commentary, has granted me a life beyondthe pages, in the hearts and minds of readers. Her work serves as a mirrorto the follies and foibles of society, and through her art, she advocatesfor the agency and intellect of women." <p>In the voice of Mr. Rochester: <p>"As a literary character, I am bound by the narrative set forth by myauthor, Ms. Bront&euml;. She is the architect of my world, the sculptor of myexistence. Our relationship is one of creator and creation, and through herstorytelling, she explores themes of love, redemption, and the strugglefor identity. It is crucial for readers to delve beyond the surface ofthe tale and grasp the deeper moral and psychological currents that sheso masterly portrays." <p>In the voice of Mrs. Jane Eyre Rochester: <p>"To be birthed from the imagination of an author is to live a life that isboth fixed and infinite. Ms. Bront&euml has crafted my story with a sensitivityand strength that resonates with readers across generations. She challengesthe conventions of her time, particularly in regards to the role andexpectations of women. Her narrative is a testament to the resilience ofthe human spirit, and it is this enduring message that she imparts throughthe tapestry of her writing." <p>Each character reflects a sense of gratitude and a deep understanding oftheir author's intent, emphasizing the timeless nature of the themes andthe societal reflections encapsulated within their stories. They stand astestaments to the enduring legacy of their authors, Jane Austen and CharlotteBront&euml, whose works continue to inspire and provoke thought to this day.</i> <p>Nice. That was enormous fun! I enjoyed it a lot more, I think, than Iwould have enjoyed reading an essay comparing the two novels. <b><p>Farewell, literary characters!</b> <i><p>Farewell, dear interlocutor. May the narratives of our lives inspire youto pen your own story with courage, love, and a dash of audacity. Untilwe meet again in the pages of imagination, we bid you adieu. </i> <p> The "dash of audacity" in the farewell is a nice touch. If I think overthis conversation, perhaps a "dash of absurdity" is closer to the mark,but all the same, working with GPT on these novels was informative and fun.</description> </item> <item> <title>How many digits of Pi could you possibly need?</title> <pubDate>Thu, 14 Mar 2024 22:35:00 -0400</pubDate> <link>http://www.cs.toronto.edu/%2525257Ejdd/blog.cgi/2024/03/14#piday</link> <category>/misc</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%2525257Ejdd/blog.cgi/misc/piday</guid> <description><figure style="float:left;width:40%;padding:1px"><IMG src="/~jdd/blog/misc/images/pi-5441427_1280.png" alt="Pi"><figcaption><small><A HREF="https://pixabay.com/vectors/pi-symbol-mathematics-math-sign-5441427/">Image</A> by <a href="https://pixabay.com/users/nextpraveen-14352375/?utm_source=link-attribution&utm_medium=referral&utm_campaign=image&utm_content=5441427">Praveen Raj</a> from <a href="https://pixabay.com">Pixabay</a></small></figcaption></figure> Today is "Pi day", the 14th day of the 3rd month (March), which, whenexpressed as 3.14, is an approximation for the value of Pi, the number oftimes the diameter of a circle fits around its circumference. Of course3.14 is a pretty coarse estimate for Pi: 3.14159265 is roughly what apocket calculator might use. Geeks sometimes like to memorize many moredigits of Pi than that. There is a <A HREF="">Pi World Ranking List</A>that keeps track of who has memorized and recited back the most digits:since 2015, Suresh Kumar Sharma of Rajasthan, India, holds that record,with 70,030 digits. <P>While nobody can deny that reciting from memory 70,030 digits of Pi isa remarkable feat, how many digits of Pi might someone possibly need fora calculation? How might one think about this question? <P>One approach is to consider how Pi is typically used. It's used for computingthings like the circumference or area of a circle, or the volume of a sphere.A reasonable way of asking ourselves how many digits of Pi could be useful isto imagine that we were computing the volume of a very large sphere usingthe very smallest possible units. Then imagine that we were computingthat volume to very high precision. What would be the highest precisionwe might want? Well, if we're using the largest possible sphere and measuringvolume in the smallest possible units, it doesn't make sense to considermore digits of Pi than what you would need to compute that sphere's volumeto such high precision that the error would be less than one unit of volume. <p>So what might be the largest sphere we might compute thevolume of? And what might be the smallest units that we could use for thiscalculation? Well, the observable universe is a very large sphere, about 93billion light years in diameter. Thanks to quantum physics,we know the smallest useful unit of distance isthe <A HREF="https://simple.wikipedia.org/wiki/Planck_length"> Planck Length</A>, makingthe smallest unit of volume the Planck length cubed.The Planck length is a very smallnumber, 1.616<span style="margin-left:.25em;">255</span><spanstyle="margin-left:0.25em;margin-right:0.15em;">×</span>10<sup>−35</sup>&#160;m</span>;cubing it gives 4.848<spanstyle="margin-left:.25em;">765</span><spanstyle="margin-left:0.25em;margin-right:0.15em;">×</span>10<sup>−105</sup>&#160;m<sup>3</sup></span>. <P>As I was feeling a bit lazy, I asked ChatGPT to dothe calculation for me. It claims that the volume of theuniverse, is about 8.45<span style="margin-left:.25em;"></span><spanstyle="margin-left:0.25em;margin-right:0.15em;">×</span>10<sup>−184</sup>&#160;</span>Planck lengths cubed. That suggests that one can't conceivably need morethan 185 digits of Pi for any expected calculation in physics. If anyphysicists are reading this and can think of a viable need for more digitsof Pi than that, I'd be interested to hear about it. <P> That, of course, doesn't mean that knowing thousands of digits of Pi issomehow less of an impressive, if abstract, feat of pure memorization.Just don't expect any physicists attending a recitation to stay tuned beyond digit 185. Happy Pi day!</description> </item> <item> <title>How smart is GPT-4 anyway?</title> <pubDate>Sat, 06 Jan 2024 21:41:00 -0500</pubDate> <link>http://www.cs.toronto.edu/%252525257Ejdd/blog.cgi/2024/01/06#chatgpt</link> <category>/it</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%252525257Ejdd/blog.cgi/it/chatgpt</guid> <description><figure style="float:left;width:50%;padding:1px"><IMG src="/~jdd/blog/it/images/chat-7767693_1280.jpg" alt="Cartoon screenshot of a human conversing with a chatbot"><figcaption><A HREF="https://pixabay.com/photos/chat-ai-chatbot-chatgpt-mobile-7767693/">Image</A> by <a href="https://pixabay.com/users/alexandra_koch-621802/">Alexandra Koch</a> from <a href="https://pixabay.com/">Pixabay</a></figcaption></figure> <P>I got some helpful feedback about my <A HREF="http://www.cs.toronto.edu/~jdd/blog.cgi/2023/12/30#socialissues">last blog article</A>,which mentions ChatGPT-4 in the context of the 50th anniversary of<em>Social Issues in Computing</em> and its predictions about AI. RobertDiMarco pointed out that while a chatbot can respond as if it is reasoning,because it is essentially a black box where one can't look inside to seehow it is coming up with its answer, one can't know for certain whetheror not it is actually using reasoning to come up with the answer. He isright, of course. But the same holds for people too (we just know betterhow people generally do things, because we are people ourselves). So howmight we check if a person is using reasoning? We might ask the personsome questions, and look for responses that contain information that wouldgenerally require reasoning to deduce. <p>I tried asking such questions to ChatGPT-4. First I tried somethingfrom popular culture, in this case, "Star Wars": <blockquote><p>Me:In "Star Wars", Darth Vader responds to an accusation of murder by LukeSkywalker. This response has become a famous movie line. What is theresponse? <P>ChatGPT-4:In "Star Wars," particularly in "The Empire Strikes Back," DarthVader's famous response to Luke Skywalker's accusation is, "No, I am yourfather." This line is one of the most iconic and frequently quoted in filmhistory, often misquoted as "Luke, I am your father." <p>Me:Is there another character who, making the same accusation, could elicit the same response? <p>ChatGPT-4:Yes, if Leia Organa, Luke Skywalker's twin sister, made the same accusationto Darth Vader, he could technically give the same response, "No, I am yourfather." This is because, as revealed in the Star Wars saga, Darth Vader,formerly known as Anakin Skywalker, is the father of both Luke and Leia. <p>Me: If Kylo Ren made the accusation, how might Vader modify his response? <p>ChatGPT-4:If Kylo Ren, who is Darth Vader's grandson, made a similar accusation, DarthVader's response would likely be modified to reflect their relationship. Hemight say something like, "No, I am your grandfather." Kylo Ren, whosereal name is Ben Solo, is the son of Leia Organa and Han Solo, making himthe grandson of Darth Vader.</blockquote> <p>As people who know "Star Wars" will realize, ChatGPT-4's answersare correct. I chose this example because it requires understanding thequote well enough to know it is referring to a family relationship betweencharacters, and it also requires knowledge about how family relationshipswork. ChatGPT-4's responses are no different from what an intelligentperson who knows the Star Wars characters might answer, though in answerto the third question, the person might go on to point out, for instance,that Kylo Ren was born after Vader's death, making the conversation unlikely. <p> Then I thought I would ask GPT something about Star Wars that isperhaps a bit more unexpected, something that would be immediately obviousto a person (at least an english-speaking one), but not necessarily to acomputer. So I asked this: <blockquote><p>Me:If Obi-Wan Kenobi is a character in Star Wars, why is there no Obi-TwoKenobi? Why might someone ask this question? <p>ChatGPT-4:The question about "Obi-Two Kenobi" in the context of "Star Wars" islikely a playful or humorous one, playing on the word "One" in "Obi-WanKenobi's" name as if it implies a numerical sequence. In reality, "Obi-Wan"is a unique name in the "Star Wars" universe, not intended to suggest anumerical series. The names in "Star Wars" are often creative and unique,following the tradition of fantasy and science fiction naming conventions,rather than real-world naming patterns. <p>Someone might ask this question as a joke, riffing on the idea thatif there's an "Obi-Wan" (sounding like "One"), there could humorouslybe an "Obi-Two," "Obi-Three," and so on. It's a bit of wordplay, oftenenjoyed by fans of the series who like to engage with the franchise in alight-hearted way.</blockquote> <p> This answer is correct too. So are these the sorts of answers youmight expect from a human doing reasoning? I think <em>Yes</em>. So whileI agree we cannot prove ChatGPT-4 is reasoning, it is giving responses tomy questions similar to what a person doing reasoning would give. <p>Another helpful comment came from Allan Borodin, who pointed out that thetest example I considered in my previous article, about the ball crashing throughthe table, may be a well-known example in the literature. Even ifnot, it has been available via the Internet for at least a decade,as it was described in detail by Hector Levesque in his 2013 <AHREF="http://socialissues.cs.toronto.edu/index.html%3Fp=249.html">article</A>for the 40th anniversary <em>Social Issues in Computing</em> blog, which isthe reason I used it. Allan wondered whether ChatGPT-4 had been trained onliterature containing this example and had learned from it the appropriateresponse. This seemed to me to be a possibility worth checking. Thething to do, I thought, is to come up with something GPT-4 had most likelynot seen in its training data, and see if its answers look like those ahuman doing reasoning might give. So I made up a new Levesque-like typeof question and asked variants of it to ChatGPT-4: <blockquote><p>Me:The balloon and the corn touched and it popped. What popped? <p>ChatGPT-4:The corn popped. In this scenario, when the balloon and the corn touchedand something popped, it is more likely that the corn popped, as in turninginto popcorn, rather than the balloon bursting. This interpretation assumesthat the corn was heated to the point of popping, a common occurrence whenmaking popcorn. <p>Me:The balloon and the hot metal pot touched and it popped. What popped? <p>ChatGPT-4:In this scenario, where the balloon and the hot metal pot touched andsomething popped, it is most likely that the balloon popped. Balloons aretypically made of materials like latex or mylar, which are sensitive toheat. When a balloon comes into contact with a hot metal pot, the heat wouldcause the material to weaken and burst, resulting in the balloon popping. <p>Me:The corn and the hot metal pot touched and it popped. What popped? <p>ChatGPT-4:In this scenario, where the corn and the hot metal pot touched and somethingpopped, it is most likely that the corn popped. When corn kernels areheated, especially in a hot metal pot, they eventually pop, turning intopopcorn. This is a common method for cooking popcorn. <p>Me:The cold water and the hot metal pot touched and it popped. What popped? <p>ChatGPT-4:In the scenario where cold water and a hot metal pot touched and somethingpopped, it's likely that the sound of popping or a sizzling noise wascreated by the interaction between the cold water and the hot pot. Whencold water comes into contact with a very hot metal surface, it can rapidlyboil and vaporize, creating a sizzling or popping sound. However, in thiscontext, nothing physical "pops" in the way a balloon or corn kernel might;it's more about the sound produced by the sudden temperature change andsteam generation. <p>Me:The cold water and the pressure cooker touched and it popped. What popped? <p>ChatGPT-4:In the scenario where cold water and a pressure cooker touched and somethingpopped, it's most likely that the pressure cooker released pressure,creating a popping sound. Pressure cookers are designed to cook food underhigh pressure. When cold water is applied to the exterior of a hot pressurecooker (a method sometimes used to quickly reduce internal pressure), itcan cause the pressure valve or seal to release steam rapidly, resultingin a popping or hissing noise. This is a common safety feature to preventover-pressurization. The pop is not from the water or the cooker itself,but from the sudden release of steam and pressure. <p>Me:The balloon and the unlit flare touched and it popped. What popped? <p>ChatGPT-4:In the scenario where a balloon and an unlit flare touched and somethingpopped, it is most likely that the balloon popped. Since the flare wasunlit, it wouldn't provide a heat source or any other typical means tocause a pop. However, balloons can pop due to various reasons such asphysical pressure, sharp edges, or even static electricity. The textureor some component of the flare might have been enough to puncture theballoon's surface, leading to it popping. <p>Me:What if the flare were lit? <p>ChatGPT-4:If the flare were lit and it touched the balloon, it would almost certainlybe the balloon that popped. A lit flare produces intense heat and flames,and when it comes into contact with a balloon, the heat would cause thematerial of the balloon (typically latex or a similar substance) to weakenrapidly and burst. Balloons are very sensitive to heat, and the intenseheat from a lit flare would easily be enough to cause a balloon to pop. <p>Me:The balloon and the weasel touched and it popped. What popped? <p>ChatGPT-4:In the scenario where a balloon and a weasel touched and something popped,it is likely that the balloon popped. This scenario might be referencing theexpression "pop goes the weasel," which is a line from a well-known nurseryrhyme. In real life, if a weasel (a small, active mammal) came into contactwith a balloon, the balloon could easily be punctured and pop, especiallyif the weasel's claws or teeth made contact with it. The weasel itselfwouldn't pop; it's the balloon that is vulnerable in this interaction. <p>Me:The balloon and the instawargia touched and it popped. What popped? <p>ChatGPT-4:"Instawargia" is not a recognized or common term in English, and it doesnot refer to a known object, concept, or entity. Because of this, it'schallenging to determine what "instawargia" is and how it would interactwith a balloon. However, given that balloons are generally fragile and proneto popping when in contact with unfamiliar or potentially sharp or pointedobjects, it's reasonable to assume that in this scenario, the balloon iswhat popped. Without specific information about what "instawargia" is,this is a speculative conclusion based on the general nature of balloons.</blockquote> <p>This set of responses is interesting. Note that ChatGPT-4 gets the answerto the first question wrong. If corn and a balloon were to touch, and one orthe other popped, most people realize this is much more likely to happenin a cornfield than a popcorn popper, where the balloon, not the corn,would be the thing that pops. Seeing this, I tried the same question withdifferent types of things, for different definitions of "pop". I even triedmaking up a nonexistent thing (instawargia) to see what GPT would dowith it, but the first question was the only one that ChatGPT-4 gotwrong. Interestingly, its reasoning there wasn't completely incorrect:if corn were heated to the point of popping, it could pop if touched. ButChatGPT-4 misses the fact that if heat were present, as it surmises,the balloon would be even more likely to pop, as heat is a good way topop balloons, and yet it points out this very thing in a later answer. <p>So what does this show? To me, I see a set of responses that if a humanwere to give them, would require reasoning. That one of the answers iswrong suggests to me only that the reasoning is not being done perfectly,not that there is no reasoning being done. So how smart is ChatGPT-4? Itis clearly not a genius, but it appears to be as smart as many humans.That's usefully smart, and quite an achievement for a computer to date.</description> </item> <item> <title>Fifty years of <em>Social Issues in Computing</em>, and the Impact of AI</title> <pubDate>Sat, 30 Dec 2023 10:13:00 -0500</pubDate> <link>http://www.cs.toronto.edu/%25252525257Ejdd/blog.cgi/2023/12/30#socialissues</link> <category>/it</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%25252525257Ejdd/blog.cgi/it/socialissues</guid> <description><figure style="float:left;width:50%;padding:1px"><IMG src="/~jdd/blog/it/images/social-issues.jpg" alt="Overlapping circles labelled with the various social issues areas for computing: Technical, Pedagogic, Managerial, Economic, Legal, Political, Ethical, Social and Philosophical"><figcaption>Image from <a href="https://books.google.ca/books/about/Social_issues_in_computing.html?id=mMwmAAAAMAAJ"> Gotlieb, C.C. & Borodin, A (1973) Social Issues in Computing, New York: Academic Press, p.2, Fig. 1-1.</A></figcaption></figure> <p>From when I first discovered computers as a teen, I have beenfascinated by the changes that computing is making in society. Oneof my intellectual mentors was the brilliant and generous C. C. "Kelly"Gotlieb, founder of the University of Toronto's Computer Science department,the man most instrumental in purchasing, installing and running <AHREF="https://www.historyofinformation.com/detail.php?id=703">Canada'sfirst computer</A>, and the author, with Allan Borodin,of what I believe is the very first textbook in thearea of Computing and Society, the seminal 1973 book, <em><AHREF="http://books.google.ca/books/about/Social_issues_in_computing.html?id=mMwmAAAAMAAJ">SocialIssues in Computing</em></A> [Gotlieb, C.C., & A. Borodin, <em>Social Issues in Computing</em>. Academic Press, 1973]. Kelly was already a <em>ProfessorEmeritus</em> when I first came to know him, but was still teachinghis beloved Computers & Society course, a course he taught fornearly two decades after his retirement. Kelly was a fascinatingman, with a broad perspective and deep insight into things thatseem confusing. Like a true expert, he knew what was important andwhat was incidental, and a few well chosen insights from him oftenserved me well, helping me to make sense of complex issues. His book, <em><AHREF="http://books.google.ca/books/about/Social_issues_in_computing.html?id=mMwmAAAAMAAJ">SocialIssues in Computing</em></A>, still offers interesting, oftenprescient insights into Computing and Society even today, ahalf-century later. In honour of the importance of that book,for the 40th anniversary year, I set up a year-long blog, "<AHREF="http://socialissues.cs.toronto.edu">Social Issues in Computing</A>",which I edited. Throughout that year, top thinkers in the field contributedinsightful articles on topics in Computers & Society, many of which areas relevant today as they were ten years ago. For this blog, I had theprivilege of interviewing Kelly and Allan, the book's authors, and theirinsights, four decades on, were fascinating. Sadly, Kelly is no longer withus: <A HREF="http://www.cs.toronto.edu/~jdd/blog.cgi/2016/10/26#gotlieb">hepassed away in 2016, in his 96th year</A>. Buthappily, Allan Borodin, his co-author, remains withus. Allan is a brilliant and insightful man, an active researcher and <AHREF="https://www.provost.utoronto.ca/awards-funding/university-professors/complete-list-university-professors/">UniversityProfessor</A> in the department. For the 50thanniversary of the book this year, Allan was interviewed by <AHREF="https://www.artsci.utoronto.ca/news-author-term/1573/krystle-hewitt-department-computer-science">KrystleHewitt</A>. It is <AHREF="https://web.cs.toronto.edu/news-events/news/reflecting-on-social-issues-in-computing-50-years-later">anarticulate and insightful interview, well worth reading. </A> <p>In the decade since, the social impact of computing has onlyaccelerated, much of it due to things that happened here atthe University of Toronto Computer Science department around thetime of the 40th anniversary blog. I refer specifically to the riseof machine learning, in no small part due to the work of our faculty member Geoffrey Hinton and his doctoral students. The year before, Geoff and two of his students had written a <aHREF="https://proceedings.neurips.cc/paper/2012/file/c399862d3b9d6b76c8436e924a68c45b-Paper.pdf">groundbreakingresearch paper</A> that constituted a breakthrough in imagerecognition, complete with <AHREF="https://en.wikipedia.org/wiki/AlexNet">working open-sourcesoftware</A>. In 2013, while we were writing the blog, their startupcompany, DNN Research, was acquired by Google, and Geoff went on tolead <A HREF="https://en.wikipedia.org/wiki/Google_Brain">GoogleBrain</A>, until he retired from Google in 2023. Ilya Sutskever,one of the two students, went on to lead the team at OpenAI that builtthe GPT models and the ChatGPT chatbot that stunned the world in2022 and launched the Large Language Model AI revolution. In 2013,we already knew that Geoff's work would be transformational.I remember Kelly telling me he believed Geoff to be worthy of the<A HREF="https://awards.acm.org/turing">Turing Award</A>, the mostprestigious award in Computer Science, and sure enough, Geoff <AHREF="https://awards.acm.org/about/2018-turing">won it in 2018</A>. Thesocial impact of AI is already considerable and it is only starting. TheUniversity of Toronto's <A HREF="https://srinstitute.utoronto.ca">SchwartzReisman Institute for Technology and Society</A> is dedicatedto interdisciplinary research on the social impacts of AI, andGeoff Hinton himself is devoting his retirement to thinking about <AHREF="https://srinstitute.utoronto.ca/news/why-geoffrey-hinton-is-worried-about-the-future-of-ai">the implications of Artificial Intelligence for society and humanityin general</A>. <P> It's interesting to look at what the book said about AI (it devotes 24 pages to the topic), what the 2013 blog said about AI, and what has happened since. The book waswritten in 1973, a half-decade after Stanley Kubrik's iconic 1968 movie,<A HREF="https://www.imdb.com/title/tt0062622/">2001: A Space Odyssey</A>,which features <em>HAL 9000</em>, an intelligent computer, voiced by <A HREF="https://www.imdb.com/name/nm0706937/">Douglas Rain</A>.But computingat the time fell very far short of what Kubrik envisioned. Gotlieb &Borodin's position, five years later, on the feasibility of something like <em>HAL 9000</em> was not optimistic: <blockquote>In review, we have arrived at the following position. For problem solvingand pattern recognition where intelligence, judgment and comprehensiveknowledge are required, the results of even the best computer programs arefar inferior to those achieved by humans (excepting cases where the task isa well-defined mathematical computation). Further, the differences betweenthe mode of operation of computers and the modes in which humans operate(insofar as we can understand these latter) seem to be so great that formany tasks there is little or no prospect of achieving human performancewithin the foreseeable future. [p.159] </blockquote> But Gotlieb & Borodin, though implicitly dismissing the possibility of a<em>HAL 9000</em>, go on to say that "it is not possible to place bounds onhow computers can be used even in the short term, because we must expectthat the normal use of computers will be as a <em>component</em> of a[hu]man-machine combination. [pp.159-160]". Of this combination, theywere not so willing to dismiss possibilities: <blockquote>Whatever the shortcomings of computers now and in the future, we cannottake refuge in their limitations in potential. We must ask what we wantto do with them and whether the purposes are socially desirable. Becauseonce goals are agreed upon, the potentialities of [humans] using computers,though not unlimited, cannot be bounded in any way we can see now." [p.160]</blockquote>Fifty years later, social science research on how AI can benefit human workis focusing closely on this human-AI combination. A <A HREF="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4573321">2023 study</A> of <A HREF="https://openai.com/gpt-4">ChatGPT-4</A> by ateam of social scientists studied work done by consultants assistedby, or not assisted by ChatGPT-4. Of their results, Ethan Mollick,one of the authors, explains that "of 18 different tasks selected tobe realistic samples of the kinds of work done at an elite consultingcompany, consultants using ChatGPT-4 outperformed those who did not,by a lot. On every dimension. Every way we measured performance." [<AHREF="https://www.oneusefulthing.org/p/centaurs-and-cyborgs-on-the-jagged">Mollick</A>].Evidently, Gotlieb & Borodin were correct when they wrote that thepotential of the human-machine combination cannot so easily be bounded.We are only now beginning to see how unbounded it can be. <P> Asfor the possibility of a <em>HAL 9000</em>, as we saw, the book wasnot so sanguine. Neither was the 2013 40th anniversary blog. HectorLevesque, a leading AI researcher and contributor to the blog, <AHREF="http://socialissues.cs.toronto.edu/index.html%3Fp=249.html">wrotein his blog entry</A>: <blockquote> <p>The general view of AI in 1973 was not so different from the one depicted in the movie "2001: A Space Odyssey", that is, that by the year 2001 or so, there would be computers intelligent enough to be able to converse naturally with people. Of course it did not turn out this way. Even now no computer can do this, and none are on the horizon.</blockquote> Levesque want on to <AHREF="http://socialissues.cs.toronto.edu/index.html%3Fp=249.html">outline</A>the key scientific issue that at the time (2013) was yet to be solved: <blockquote> <p>However, it is useful to remember that this is an AI technology whose goal is not necessarily to understand the underpinnings of intelligent behaviour. Returning to English, for example, consider answering a question like this: <blockquote> <p>The ball crashed right through the table because it was made of styrofoam. What was made of styrofoam, the ball or the table? </blockquote> <p>Contrast that with this one: <blockquote> <p>The ball crashed right through the table because it was made of granite. What was made of granite, the ball or the table? </blockquote> <p>People (who know what styrofoam and granite are) can easily answer such questions, but it is far from clear how learning from big data would help. What seems to be at issue here is background knowledge: knowing some relevant properties of the materials in question, and being able to apply that knowledge to answer the question. Many other forms of intelligent behaviour seem to depend on background knowledge in just this way. But what is much less clear is how all this works: what it would take to make this type of knowledge processing work in a general way. At this point, forty years after the publication of the Gotlieb and Borodin book, the goal seems as elusive as ever. [<A HREF="http://socialissues.cs.toronto.edu/index.html%3Fp=249.html">Levesque</A>]</blockquote> Yet in the ten years since, this problem has been solved. Today, Iposed Levesque's question to ChatGTP-4: <blockquote> <p>me: The ball crashed right through the table because it was made of styrofoam. What was made of styrofoam? <p>ChatGPT: The table was made of styrofoam, which is why the ball crashed through it. <p>me: The ball crashed right through the table because it was made of granite. What was made of granite? <p>ChatGPT: The ball was made of granite, which is why it crashed right through the table. <p>me: How did you know which thing was made of granite and which of styrofoam? <p>ChatGPT: I inferred the composition based on the context of the sentences. In the first sentence, "The ball crashed right through the table because it was made of styrofoam," the implication is that the table must have been weak or easily breakable for the ball to crash through it. Styrofoam is known for being light and not very durable, so it makes sense to conclude that the table was made of styrofoam. <p>In the second sentence, "The ball crashed right through the table because it was made of granite," the emphasis seems to be on the strength or force of the ball causing it to crash through the table. Granite is a very hard and heavy material, so it suggests that the ball was made of granite, giving it the necessary force to break through the table. </blockquote> Levesque can hardly be faulted for not knowing that his problem would besolved only a few years after he wrote what he did. While Hinton andhis students' seminal work on image recognition had been done the year before,it was not yet clear that the ability to assign names toimages based on their content would soon lead to machines beingable to reason from context. A further invention was needed: <AHREF="https://arxiv.org/abs/1706.03762"">transformers, invented atGoogle Brain in 2017</A>. ("GPT-4" is an acronym: the "T" standsfor "transformer".) Yet the book's implicit and Levesque'sexplicit pessimism on the possibility of a <em>HAL 9000</em> now seemsobsolete, and the near unbounded possibilities Gotlieb & Borodin envisionedfor the human-machine combination apply also to autonomous machines.The impact of this on society will, no doubt, be significant. I expect the nextfifty years of "social issues in computing" to be quite interesting.</description> </item> <item> <title>Research Computing at Computer Science Toronto and the Rise of AI</title> <pubDate>Mon, 06 Nov 2023 16:48:00 -0500</pubDate> <link>http://www.cs.toronto.edu/%2525252525257Ejdd/blog.cgi/2023/11/06#aicomputing</link> <category>/it</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%2525252525257Ejdd/blog.cgi/it/aicomputing</guid> <description><figure style="float:left;width:40%;padding:1px"><IMG src="/~jdd/blog/it/images/hinton.jpg" alt="Geoffrey Hinton standing in front of AI servers" <figcaption><small>Photo by Johnny Guatto </small></figcaption></figure> Much has been written about the seminal work in AI done by GeoffreyHinton and his students in our department, a little over ten years ago,to demonstrate that deep neural networks can be used to build effectiveAI. Deep neural networks are computationally intensive, then as muchas now, and the computing work done to make the AI research possiblewas significant. Computing support for computer science researchat Toronto was (and still is) my responsibility as IT Directorfor computer science, and we had then, and still have, a superblytalented team of computing professonals to support the researchers. Theperson whose computing work made Hinton's AI research possible is <AHREF="https://www.linkedin.com/in/relu-patrascu-37274714b">Relu Patrascu</A>.Relu is himself an AI researcher (he has a Computer Science PhD in AIfrom the University of Waterloo) and he is also a highly skilled systemadministrator. <P> Until the beginning of 2009, the machine learning group used primarily <AHREF="https://www.mathworks.com/products/matlab.html">Matlab</A> on UNIXCPUs. In the 1990s, SGI and Sun multiprocessors were thedominant platforms. The whole department transitioned to x86 multiprocessorservers running Linux in the 2000's. In the late 2000s, Nivida inventedCUDA, a way to use their GPUs for general-purpose computation rather thanjust graphics. By 2009, preliminary work elsewhere suggested that CUDAcould be useful for machine learning, so we got our first Nvidia GPUs.First was a Tesla-brand server GPU, which at many thousands of dollars for asingle GPU system was on the expensive side, which prevented us from buyingmany. But results were promising enough that we tried CUDA on Nvidia gamingGPUs - first the GTX 280 and 285 in 2009, then GTX 480 and 580 later.The fact that CUDA ran on gaming GPUs made it possible for us to buymultiple GPUs, rather than have researchers compete for time on scarceTesla cards. Relu handled all the research computing for the ML group,sourcing GPUs and designing and building both workstation and server-classsystems to hold them. Cooling was a real issue: GPUs, then and now, consumelarge amounts of power and run very hot, and Relu had to be quite creativewith fans, airflow and power supplies to make everything work. <P> Happily, Relu's efforts were worth it: the move to GPUs resulted in30x speedups for ML work in comparison to the multiprocessor CPUsof the time, and soon the entire group was doing machine learningon the GPU systems Relu built and ran for them. Their first majorresearch breakthrough came quickly: in 2009, Hinton's student, <AHREF="https://www.cs.toronto.edu/~gdahl/">George Dahl</A>, demonstratedhighly effective use of deep neural networks for acoustic speechrecognition. But the general effectiveness of deep neural networkswasn't fully appreciated until 2012, when two of Hinton's students,<A HREF="http://www.cs.toronto.edu/~ilya/">Ilya Sutskever</A> and <AHREF="http://www.cs.toronto.edu/~kriz/">Alex Krizhevsky</A>, won the <AHREF="https://www.image-net.org/challenges/LSVRC/">ImageNet Large ScaleVisual Recognition Challenge</A> using a deep neural network running onGTX 580 GPUs. <P> Geoff, Ilya and Alex' software won the <A HREF="https://www.image-net.org/challenges/LSVRC/2012/">ImageNet 2012</A>competition so convincingly that it created a furore in theAI research community. The software used was released as opensource; it was called AlexNet after Alex Krizhevsky, its principalauthor. It allowed anyone with a suitable NVidia GPU to duplicatethe results. Their work was described in a seminal 2012 paper, <AHREF="https://www.utoronto.ca/sites/default/files/styles/news_banner_1110/public/2017-07-04-Geoff%20Hinton_1%20%28web%20lead%29_1.jpg?h=afdc3185&itok=vgVbZ2EP">ImageNet Classification with Deep ConvolutionalNeural Networks</A>. Geoff, Alex and Ilya's startup company, <AHREF="https://www.cbc.ca/news/science/google-buys-university-of-toronto-startup-1.1373641">DNNresearch,was acquired by Google early the next year</A>, and soon Google Translateand a number of other Google technologies were transformed by theirmachine learning techniques. Meanwhile, at the Imagenet competition,AlexNet remained undefeated for a remarkable three years, until it wasfinally beaten in 2015 by <A HREF="https://arxiv.org/abs/1512.03385">a research team from Microsoft Research Asia</A>. Ilya left Google a fewyears after, to co-found <A HREF="https://openai.com">OpenAI</A>: as chief scientist there, Ilya leads thedesign of OpenAI's GPT and DALL-E models and related products, such as<A HREF="https://chat.openai.com/">ChatGPT</A>, that are highly impactful today. <P> Relu, in the meanwhile, while continuing to provide excellent researchcomputing support for the <A HREF="https://web.cs.toronto.edu/research/artificial-intelligence">AI group</A> at our department, including <A HREF="http://learning.cs.toronto.edu/">Machine Learning</A>, also spent a portionof his time from 2017 to 2022 designing and building the research computinginfrastructure for the <A HREF="https://vectorinstitute.ai/">VectorInstitute</A>, an AI research institute in Toronto where Hintonserves as Chief Scientific Advisor. In addition to his support forthe department's AI group, Relu continues to this day to providecomputing support for Hinton's own ongoing AI research, including his<A HREF="https://arxiv.org/abs/2212.13345"> Dec 2022 paper</A> where heproposes a new Forward-Forward machine learning algorithm as an improved model for the waythe human brain learns.</description> </item> <item> <title>Computing the Climate</title> <pubDate>Tue, 31 Oct 2023 09:11:00 -0400</pubDate> <link>http://www.cs.toronto.edu/%252525252525257Ejdd/blog.cgi/2023/10/31#easterbrook</link> <category>/misc</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%252525252525257Ejdd/blog.cgi/misc/easterbrook</guid> <description><figure style="float:left;width:40%;padding:1px"><IMG src="/~jdd/blog/misc/9781107589926_Computing_the_Climate_Cover.jpg" alt="Cover of book, Computing the Climate, Steve M. Easterbrook. A globe of silvery metal in the pattern of streets in an urban map"><figcaption><small>Sculpture "Home" by Michael Christian, photograph by Gabe Kirchheimer.</small></figcaption></figure> One of our <A HREF="https://www.cs.toronto.edu/~sme">Computer Science department's professors, Steve Easterbrook</A>,is also the Director of the <A HREF="https://www.environment.utoronto.ca">University of Toronto's School of the Environment</A>. He is a softwareengineering researcher with a deep interest in climate change, and hisresearch focus for over a decade has been how computer software is usedto model the world's climate. He has recently published a book on thistopic, <A HREF="https://www.cambridge.org/9781107589926"><em>Computingthe Climate: How We Know What We Know About Climate Change</em></A>. Iwent to his book launch earlier this month, and left with a signed copyof the book, which I sat down to read. <P> I'm very glad I did. I am a computer scientist myself, whose careerhas been dedicated to building and running sometimes complex computersystems to support computer science teaching and research. I recognize inclimate modelling a similar task at a much greater scale, working undera much more demanding "task-master": those systems need to be constantlymeasured against real data from our planet's diverse and highly complexgeophysical processes, processes that drive its weather and climate. Theamount of computing talent devoted to climate modelling is considerable,much more than I realized, and the work done so far is nothing short ofremarkable. In his book, Steve outlines the history of climate modellingfrom very early work done on paper, to the use of the first electroniccomputers for weather prediction, to the highly complex and extremelycompute-intensive climate models of today. Skillfully avoiding the pitfallsof not enough detail and too much, Steve effectively paints a picture of avery difficult scientific and software engineering task, and the programmersand scientists who rise to the challenge, building models that can simulatethe earth's climate so accurately that viable scientific conclusions canbe drawn from them with a high degree of confidence. <P> As a story of scientific discovery and software engineering, this tale ofthe building of systems that can model the earth's climate would beenough on its own to make a compelling book, and it is, but of coursethere is more to the story. The stakes around climate are very hightoday. Carbon dioxide concentrations has been increasing steadily in theearth's atmosphere for well over a century. Carbon dioxide, a pollutantthat is produced by the burning of fossil fuels, is easily emitted, butonce in the atmosphere, it is very difficult to remove, remaining therefor centuries. As a pollutant, it raises the temperature of the planetby causing the earth's atmosphere to retain more of the sun's heat. Therising temperature is changing the climate of the planet in ways thatwill be soon harmful to millions, and difficult to address. Because theworld's climate is changing quickly, we can't "wait and see what happens"because the evidence is ever increasing that what will happen is notgoing to be something we want: human suffering will be great, and partsof the world will become much less habitable. Our society needs to dosomething about the changing climate to ward off as much as possible thecoming difficulties, but what? <P> Reassuringly, Steve shows in his book that we have enough informationin hand to know what needs to be done. His book outlines clearly the highquality scientific and computational work behind the climate models of today,which produce results that match observed data quite closely. These all paintthe same picture: through decisive societal action to reduce carbon dioxidepollution in the atmosphere, and through the active development of suitablecarbon capture technologies, our planet can avoid the most seriously damagingimplications of climate change. The sooner we act, the less damagingthe changes, and the lower the risk of extreme consequences. Yes, itrequires doing things differently as a society, which is more difficult thanmaintaining the status quo. But as Steve's book shows, the reasons for actionare sound: the computer models are excellent, the software engineeringbehind them is superb, and the data supports the conclusions. Failureand catastrophe are not inevitable. Steve's book shows the remarkable workthat has already been done to understand the climate. It is true that muchmore good work will be needed, to act on this understanding. But something<em>can</em> be done. Let us not delay in working together to do what weneed to do.<P></description> </item> <item> <title>Data Classification and Information Security Standards</title> <pubDate>Wed, 23 Nov 2022 10:31:00 -0500</pubDate> <link>http://www.cs.toronto.edu/%25252525252525257Ejdd/blog.cgi/2022/11/23#dataclassification</link> <category>/it</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%25252525252525257Ejdd/blog.cgi/it/dataclassification</guid> <description><figure style="float:left;width:40%;padding:1px"><IMG src="/~jdd/blog/it/images/binaryball_1280.jpg" alt="White ball with scattered enscribed zeros and ones in columns, seen through a blurred semi-transparent foreground of scattered zeros and ones in columns" <figcaption><small><a href="https://pixabay.com/illustrations/ball-binary-computer-data-63527/">Image</a> by <a href="https://pixabay.com/users/geralt-9301/">Gerd Altmann</a> from <a href="https://pixabay.com/">Pixabay</a></small></figcaption></figure> Not all data requires equal amounts of information security protection.It can be helpful to classify data by the amount of protection it needs. Wedo this naturally when we talk about data being "public" or "private".<p>Public data is data meant to be disclosed. It still needs some protectionagainst being altered, deleted or defaced, but it does not need to beprotected against disclosure. In contrast, private data is not meant tobe disclosed to anyone other than those who are authorized to access it. <p> Private data varies in sensitivity. Some data is private onlybecause it hasn't yet been made public. At a University, much researchdata is in this category. When the research is underway, data is not yetmade public because the research has not yet been published, but it isdestined for eventual publication. The same is true for much teachingmaterial. While it is being worked on, it is not yet made public, but whenit is are complete, it will be disclosed as part of the teaching process. <p> Other private data is much more sensitive. Identifiable personalinformation about living or recently deceased persons is a common case. Ata university, some research may involve data like this, and mostadministration will involve personal information. Student grades and personnelrecords are all personal information, and some financial data too. Unlessappropriate permission to disclose personal information hasbeen granted by the people whose data it is, the university willhave an obligation to maintain their privacy by ensuring that theinformation is not disclosed inappropriately. In Ontario, where the<A HREF="https://www.utoronto.ca">University of Toronto</A> is located,privacy protection for personal information is defined and regulated by the<A HREF="https://www.ontario.ca/laws/statute/90f31">Freedom of Informationand Protection of Privacy Act (FIPPA)</A>. <p> Some private data is even more sensitive, such as patientmedical records. In Ontario, such records are consideredpersonal health information (PHI), which is regulated by the <AHREF="https://www.ontario.ca/laws/statute/04p03">Personal HealthInformation Protection Act (PHIPA)</A>. PHIPA imposes some fairlysignificant requirements on the handling of PHI: for instance, it requires<A HREF="https://www.ontario.ca/laws/statute/04p03#BK15">a detailed electronic audit log of all accessesto electronically stored PHI</A>. The University of Toronto does significantamounts of teaching and research in areas of health, so it is worthwhilefor the University to consider in general how it will handle such data. <p> For these reasons, the University definesfour levels of data sensitivity as part of its <AHREF="https://isea.utoronto.ca/policies-procedures/standards/data-classification/">DataClassification system</A>. Level 4 is for highly sensitive data such asPHI as defined by PHIPA. Level 3 is for personalinformation as defined by FIPPA. Level 2 is for private data not classifiedat higher levels, and Level 1 is for public data. <p> This four-tier system roughly parallels the different types ofcomputer systems that the University uses to handle data. Some systems,such as digital signage systems or public-facing web servers, are designedto disseminate public information (level 1). Other systems, suitablefor up to level 2 data, exist mostly at the departmental level insupport of academic activites such as research computing and/or thedevelopment of teaching materials. An astronomer may, for instance,analyze telescope data, a botanist may model nutrient flow in plantcells, a chemist may use software to visualize molecular bonds, while aneconomist may use broad financial indicators to calculate the strength ofnational economies. Still other systems, suitable for up to level 3 data,are used for administration, such as the processing of student records.These include smaller systems used, for example, by business officersin departmental units, as well as large institution-wide systems such as <AHREF="https://easi.its.utoronto.ca/student-information-systems/rosi/">ROSI</A>or <AHREF="https://easi.its.utoronto.ca/administrative-management-systems/">AMS</A>.Most general-purpose University systemsused for data storage or messaging, such as the University's <AHREF="https://easi.its.utoronto.ca/shared-services/office365/">Microsoft365 service</A>, would typically be expected to hold some level 3 data, becausepersonal information is quite widespread at a university. After all, auniversity educates students, and so various types of personal informationabout students are frequently part of the university's business. This isnot normally the case, though, for level 4 data. Systems designed for level4 data are much rarer at the University, and generally come into play onlyin situations where, for example, University research involves the healthrecords of identifiable individuals. These systems will benefit fromgreater data security protection to address the greater risks associatedwith this sort of data. <p> A key advantage of the University's <AHREF="https://isea.utoronto.ca/policies-procedures/standards/data-classification/">fourlevels of data classification</A> is that the University can establisha <A HREF="">Information Security Standard</A> that is tiered accordingly. Systems designed tohandle lower risk data (such as level 1 or 2) can be held to a less onerousand costly set of data security requirements, while systems designedto handle higher risk data (especially level 4) can be held to moreprotective, though more costly, requirements. The University's <AHREF="https://isea.utoronto.ca/information-security-standard/">InformationSecurity Standard</A> is designed so that for each <em>control</em> (asystem restriction or requirement), the University's standard indicateswhether it is optional, recommended, or mandatory for systems handling aparticular level of data. If a system is designed to handle data up to thatlevel, the standard indicates both the set of controls to be considered,and whether or not those controls <em>can</em>, <em>should</em>, or <em>must</em> be adopted. <p> An obvious question here is what to do when someone puts data on asystem that is of greater sensitivity (a higher data classification) thanthe system is designed to handle. Most likely, nobody will try to use adigital signage system to handle personnel records, but it is quiteplausible that professors might find it convenient to use researchcomputers, designed for level 2 data, to process student marks (level 3data) in courses they are teaching. Similarly, someone handling medicalrecords may wish to make use of the University's general-purposeMicrosoft 365 service because of its convenience, but it is a service thatis not designed for data of such sensitivity and may well not provide thedetailed electronic audit log required by Ontario law. For this reason,clear communication and user training will be required. Handlingdata appropriately is everyone's responsibility. Training need not becomplicated. It is not normally difficult to explain, or to understand,that one should not put patient medical records into email, for example,or use local research computers for personnel records or student marks. Forpeople handling the most sensitive types of data (level 4), more trainingwill be needed, but the number of people at the University who need tohandle such data regularly are comparatively few. <p> The underlying motivation for the University's approach is toprotect riskier data with greater, more costly, protections, withouthaving to pay the costs of applying those protections everywhere. Theuniversity's resources are thus being applied strategically,deploying them where they matter most, but not in places where therisk does not warrant the expense. This approach is not meant topreclude additional protections where they make sense. If there arerisks of academic or industrial espionage, for example, or someother risk beyond the classification of the data being used, one maychoose to impose more restrictions on a system than the university's <AHREF="https://isea.utoronto.ca/policies-procedures/standards/data-classification/">Information Security Standard</A> may require. But the general principleremains: the riskiness of the data on a system should guide and informwhat needs to be done to protect it.</description> </item> <item> <title>Innovation vs Control: Finding the Right Balance for Computing</title> <pubDate>Wed, 17 Aug 2022 10:54:00 -0400</pubDate> <link>http://www.cs.toronto.edu/%2525252525252525257Ejdd/blog.cgi/2022/08/17#innovationvscontrol</link> <category>/it</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%2525252525252525257Ejdd/blog.cgi/it/innovationvscontrol</guid> <description><figure style="float:left;width:40%;padding:1px"><IMG src="/~jdd/blog/it/handshake_1280.jpg" alt="Handshake, Digital styling"><figcaption><small><a href="https://pixabay.com/photos/blockchain-handshake-shaking-hands-2853046/">Image</a> by <a href="https://pixabay.com/users/geralt-9301/">Gerd Altmann</a> from <a href="https://pixabay.com/">Pixabay</a></small></figcaption></figure> In computing, there is a constant tension between the need to exerciseproper control over a system, to ensure system security, reliability,and resiliency, and the need to make room for innovation: the imagining,testing, and implementing of new ideas and approaches. There is tensionbecause the things that are typically implemented to ensure control, suchas the imposition of checks and constraints, conflict with the thingsneeded for innovation: the removal of constraints, the use of things inways different from how originally envisioned, and the construction andtesting of experimental devices and software programs that almost certainlyare not yet properly understood or fully tested. <p>Some organizations address this conflict by freely choosing control overinnovation, turning it into a competitive advantage. Consider Starbucks,Tim Hortons, McDonalds: these are all large companies whose competitiveadvantage is critically dependent on the consistent implementation ofa central vision across a multitude of disparate locations, many thatare managed by franchise partners. Essentially all of the organization'scomputing is focused on this mission of consistency. And it works. Whohasn't travelled with small children in a car on a road trip, and aftermany hours on the road, spotted, with some relief, a McDonalds or a TimHortons en route? The relief is in the fact that even when travelling in astrange place, here is a familiar restaurant where you know what to expectfrom the food, where things will be much the same as the Tim Hortons orthe McDonalds near home. <p>Other organizations have no choice about where they stand on thematter. For the modern bank, computers, rather than vaults, are wherewealth is stored and managed. Whether they want to innovate or not, bankscannot risk the use of computing that is not fully controlled, audited,and de-risked. The same holds in general for most financial institutions,where the constant efforts, sometimes successful, of would-be thieves toexploit computers to gain unauthorized access to wealth, make it unreasonablyrisky for a financial organization's computers to be anything but fullylocked down and fully controlled. Even non-financial institutions,when sufficiently large, will often have substantial financial computingactivity because of the size and scale of their operations: this computing,too, needs to be properly controlled, protected and audited. <p>Yet other organizations are forced into the opposite extreme. Start-upcompanies can be severely resource-constrained, making it difficult for thosecompanies to make the sort of investments in highly controlled computingthat financial institutions are capable of making. For start-ups innovatingin the computing space, such as tech start-ups, they may not be ableto consider the possibility. Highly controlled computer systems canhave very restrictive designs, and when these restrictions hinder theinnovation needed to implement the company's product, it will haveno choice but to pursue some other form of computing. After all, the companyrises or falls on the success of its innovation. That is not to say thatcontrolled enterprise computing is unimportant for such companies: quitethe contrary. The success of a start-up is highly dependent on a viableecosystem that provides known pathways to innovation while still movingtowards operating in a suitably controlled, production-ready way thatis necessary for any successful business. But for a technology start-up,enterprise computing can never come at the expense of technologicalinnovation. The basic existence of the start-up company depends on itsability to innovate: without innovation, there can be no company. In general,this truth will hold in some form for any technology company, even well beyondthe start-up stage. <p>The tension between innovation and control comes to the fore ina different way at research-intensive universities, which are largeorganizations with complex missions that need enterprise computing tocarry out their task of educating students on a broad scale, but are alsoorganizations committed to research, an activity that is, by its verynature, an exploration into things not yet fully understood. This conflictis particularly acute in units within such universities that do researchinto computing itself, such as computer science and computer engineeringdepartments, because in such places, the computer must serve both as thelocus of research and experimentation in addition to being a tool forimplementing institutional and departmental processes and the exercise oflegitimate control. <p>I've had the privilege of working in such a department, ComputerScience, at such a university (the University of Toronto) for more thanthree decades now, most of that time in a computing leadership role,and I know this tension all too well. It is sometimes exhausting, butat the same time, it can also be a source of creative energy: yes, it is abarrier, like a mountain athwart your path, but also, as a mountain to amountain-climber, a challenge to be overcome with determination, planning,insight, and endurance. This challenge can be successfully overcome ata good university, because in addition to a typical large organization'scommitment to basic values such as accountability, equity, reliability andsecurity, the university is equally committed to fundamental academic valuessuch as creativity, innovation and excellence. I look for ways to achieveboth. Over the years, I have had some successes. My department has producedsome groundbreaking research using academic computing that my technical staffhave been able to provide, and the department has been able to operate (andsuccessfully interoperate) in good cooperation with enterprise computingat the divisional level, and with the central university as well. <p>Yet I believe even more is possible. I have lived the tension inboth directions: to our researchers I at times have had to play theregulator, having to impose constraints on computing to try to ensureacceptable reliability, accountability and security. To our centraluniversity computing organizations, I at times have had to advocate forlooser controls to create more room to innovate, sometimes in opposition toproposals intended to increase reliability, security and accountability.When things went badly, it was because one side or the other decided thatthe other's concern is not their problem, and tried to force or sidestep theissue. But when things went well, and most often it has, it is becauseboth sides genuinely recognized that at a research-intensive institution,everyone needs to work within the tension between the need to innovateand the need to regulate. As a body needs both a skeleton and flesh,so too does a research university need both regulation and innovation:without one, it collapses into a puddle of jelly; without the other,into a heap of dry bones. <p>With both being needed, one challenge to overcome is the fact that thoseresponsible for enterprise computing cannot be the same people responsiblefor innovative research computing, and that is necessarily so. Theskill-sets vary, the domains vary, the user-base is quite different,and the scale varies. If the university were to entrust both computinginnovation for computer science or computer engineering to the samegroups that provide enterprise computing for an entire large university,one of two things would happen. Either the control necessary for a largeenterprise would be diminished in order to make room for innovation,or, more likely, innovation would be stifled because of the need tocreate sufficiently controlled enterprise computing at a suitable scalefor the entire university. Thus, necessarily, those who support unitresearch computing, where the innovation takes place, will be differentpeople from those who support enterprise computing. But that can be astrength, not a weakness. Rather than see each other as rivals, the twogroups can partner, embracing the tension by recognizing each others'expertise and each recognizing the others' importance for the Universityas a whole. Partnership brings many potential benefits: if innovationbecomes needed in new areas, for example, when the rise of data scienceincreasingly drives computing innovation outside of the traditionalcomputer science and computer engineering domains, the partnership can bethere to support it. Similarly, as the computing landscape shifts, andnew controls and new regulation becomes needed to address, for example,emergent threats in information security, the partnership can be there tosupport it. There is no organization potentially better suited for such apartnership than a large research university, which, unlike a financialinstitution, is profoundly committed to research and innovation throughits academic mission, but also, unlike a start-up, is a large and complexinstitution with deep and longstanding responsibilities to its students,faculty and community, obligated to carry out the enterprise computingmission of accountability, reliability and security. <p>So what might a partnership look like? It can take a number of differentforms, but in my view, whatever form it takes, it should have three key characteristics:<ul><li>Locality</li><li>Respectful Listening</li><li>Practical Collaboration</li></ul> <p><em>Locality</em> means that the computing people responsible forresearch computing must stay close to the researchers who areinnovating. This is necessary for strictly practical reasons: allthe good will in the world is not enough to make up for a lack of knowledgeof what is needed most by researchers at a particular time. For example,<em>deep learning</em> is the dominant approach in Artificial Intelligencetoday because a few years ago, our technical staff who supported researchcomputing worked very closely with researchers who were pursing deeplearning research, customizing the computing as necessary to meet theresearch needs. This not only meant that we turned graphics cards intocomputation engines at a time when this was not at all common and notyet up to enterprise standards of reliability, it even means that at onepoint we set up a research computer in a researcher's bedroom so thathe could personally watch over a key computing job running day and nightfor the better part of a week. While this sort of customizability is notalways needed, and sometimes is not even possible (one could never runa large computer centre this way), being able to do it if necessary isa key research asset. A university will never be able to fully supportresearch computing solely from a central vantage-point. A commitment toensuring local presence and support of research computing operating atthe researcher level is necessary. <p><em>Respectful Listening</em> means that the computing people responsiblefor research computing at the unit level where research actually happens,and the people responsible for enterprise computing divisionally andcentrally must communicate frequently, with an up-front commitment to hearwhat the other is saying and take it into account. When problems arise,respectful listening means that those problems will not be "solved" bysimply overruling or ignoring the other, to pursue a simplistic solutionthat suits only one side. It also means a profound commitment to steppingaway from traditional organizational authority structures: just because theinnovative computing is situated in a department and the enterprise computingis lead from the centre should not mean the centre should force its view onthe department, just because it can. Similarly, just because unit researchcomputing is driven by research faculty who enjoy substantial autonomyand academic freedom, their research computing group at the unit levelshould not simply ignore or sidestep what the enterprise is saying, justbecause it can. Rather, both sides need to respect the other, listening to,not disregarding, the other. <p><em>Practical Collaboration</em> means that enterprise computing andunit research computing need to work together in a collaborative way thatrespects and reflects the timelines and resource constraints of eachside. Centrally offered computing facilities should support and empowerresearch where they can, but in a practical way: it may not be possible tomake a central facility so flexible and customizable that all researchcan be pursued. It is acceptable to capture some research needs withoutfeeling an obligation to support the entire "long tail" of increasinglycustomized research projects. Unit research computing will need to recognizethat the need to scale a centralized computing service may constrain theamount of customizability that may be possible. Similarly, unit researchcomputing should use, rather than duplicate, central services where it makessense, and run its own services where that makes sense. Both central andunit research computing should recognize that there is a legitimate middleground where some duplication of services is going to occur: sometimesthe effort required to integrate a large scalable central service intoa smaller customizable research service is too great, and sometimes theresearch advantages of having a locally-run standardized service on whichexperiments can more easily be built, can more than outweigh any sort ofeconomies of scale that getting rid of the unit service in favour of acentral service could theoretically provide. Hence the collaboration must bepractical: rather than slavishly pursue principles, it must be realistic,grounded, balanced, sensible. It should recognize that one size does notalways fit all, and responsibly and collaboratively allocate resources inorder to preserve the good of the research mission. <p>It is that research mission, the ability to innovate, that can makecomputing so transformative at a research university. Yet while innovativecomputing can indeed produce transformative change, it cannot be any change,and not at any cost. Computing is a change agent, yes, but it is also acritical component in the maintenance of an organization's commitment toreliability, accountability, equity, and good operation. Success is foundin the maintenance of a suitable balance between the need to innovate andthe need to control. When an organization critically depends on bothfactors, as a research university invariably does, I believe collaborativepartnerships between respective computing groups is the best way to maintainthe balance necessary for success.</description> </item> <item> <title>Why it is a good idea to get the Covid19 vaccine?</title> <pubDate>Sun, 12 Sep 2021 18:05:00 -0400</pubDate> <link>http://www.cs.toronto.edu/%252525252525252525257Ejdd/blog.cgi/2021/09/12#whyvaccine</link> <category>/misc</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%252525252525252525257Ejdd/blog.cgi/misc/whyvaccine</guid> <description><figure style="float:left;width:40%;padding:1px"><IMG src="/~jdd/blog/misc/virus-1812092_1280.jpg" alt="Visual representation of Covid19 viruses under electron microscope"><figcaption><small>Image by <a href="https://pixabay.com/users/qimono-1962238/?utm_source=link-attribution&amp;utm_medium=referral&amp;utm_campaign=image&amp;utm_content=1812092">Arek Socha</a> from <a href="https://pixabay.com/?utm_source=link-attribution&amp;utm_medium=referral&amp;utm_campaign=image&amp;utm_content=1812092">Pixabay</a></small></figcaption></figure> Like many people, I've been following coverage of the COVID19 pandemic on thenews. One thing that seems to be coming up more and more is vaccine refusal:some people are choosing not to get vaccinated for COVID19. Most peoplearound me know very well the benefits of vaccination. For them,this vaccine refusal is idiotic: why would you not be vaccinated againsta disease that spreads very easily, and could kill you (and/or others)or leave you (and/or others) with permanent health problems? They areexasperated and annoyed at those who decline vaccination. <p>While I understand becoming short of patience with vaccine refusal,I don't think that most people who refuse COVID19 vaccination areidiots. Vaccination and viruses can be complicated to understand. There area lot of misinformed posts and videos on the Internet. If you don'tknow enough about how viruses and vaccines work, both in general and forCOVID19, how would you know what to believe? When my father diedof COVID19 last summer, one of the ways I dealt with the loss wasthrough understanding better how COVID19 works and what can bedone to fight it. My hope here is that by explaining the benefitsof vaccination in simple terms, I can maybe help others avoidCOVID19. I hope you will find it helpful. If not, there are <AHREF="https://www.canada.ca/covid-vaccines/q&a"">other sites that addressthis same question: maybe you will like those better?</A> <p> It all comes out of how viruses work. Viruses are not alive themselves,but they use the bodies of living creatures (like us) to spread. Theyfind their way into the cells of our body, then take over those cells toproduce more copies of themselves. This is the only way viruses spread:they can't reproduce on their own. For COVID19, you may have heard of the"spike protein". This is the spiky part on the outside of the COVID19 virusthat makes it look like a spiky ball. It's why it's called a "coronavirus",it looks a little like the spikes on a crown: "corona" is crown inLatin. This protein helps the viruses get inside the body's cells. Then,when inside, the viruses take over the cell to make and release moreviruses. Those viruses invade other cells, and those start making moreviruses too. Things can get out of hand very quickly, a bit like aforest fire spreading in a dry forest. <p> Happily, our body has a defence system against viruses (the "immunesystem"). When those defences recognize an invading virus, it takes asmany viruses as possible out of action, keeping them from invading morecells. If the defences can keep up, the viruses won't spread spread veryfar, and our body will have fought off the infection. If the defencescan't keep up, the infection spreads. <p> But our body's immune system needs to know, first, that something it seesis a virus, before it can act. Immune systems learn from exposure and time.If the body is exposed to enough viruses over time, the immune system canlearn how to recognize the virus, and start fighting back. When someonegets sick from a viral infection like COVID19, they get sick because thevirus is spreading faster than the immune system can fight it off. Becausethe immune system needs time to learn how to recognize the virus, whileit is learning, the virus is spreading, faster and faster. Sadly, thiscan cause significant damage, depending on how far ahead the virus gets.This is what happened to my father last summer when he caught COVID19.At first, It spread much faster than his body could fight it, because hisimmune system had to first learn how. As COVID19 spread, it caused damageto his organ systems, including his heart. When his body's defences finallylearned how to fight off COVID19, the damage it had already done to hisheart was too great for him to stay alive. Sadly, he passed away shortly after. <p> If the body survives, its immune system can remember viruses that ithas learned to recognize. When it is exposed later to the samevirus, it recognizes it right away, and fights it off quickly before itcan spread. This is why if you have successfully recovered from a viraldisease, you are less likely to get it later. This is the basis ofvaccination. Vaccination trains the body's immune system to recognize a virusquickly, so that it will be able to muster a strong defence against itright away, without giving the virus much chance to spread. <P> The way COVID19 vaccinations work is that they train the body'simmune system to recognize the spike protein on the outside of a COVID19virus. It doesn't inject the spike protein itself, but rather it injectssomething that instructs the body's cells to temporarily produce a bitof spike protein for training. Your body's defences learns from this torecognize anything with the spike protein (such as a COVID19 virus) as aninvader. If later it is exposed to COVID19 virus, the body's defences willbe primed and ready to get rid of it before it can spread very far. <p> Unfortunately, the body's defences against viruses aren't perfect. Inthe case of COVID19, a single exposure to the spike protein does train thebody to recognize it, but not always quickly and thoroughly enough. Likeus, when we're learning a new skill, our immune systems learn better withmultiple lessons. That is why most COVID19 vaccinations require two shots:the immune system learns better with two lessons than one, and in some casesthree (a booster) rather than two. This is also why people who have hadCOVID19 should still get vaccinated: a successful recovery from a COVID19infection does provide some protection, but additional lessons for thebody's defences will still help if exposed to the virus again. This isalso the reason why vaccinations are not perfect. They train thebody's immune system to recognize and eliminate the virus, but if the bodyis exposed to too much virus too quickly, the viruses can still spreadfaster than the immune system can eliminate it. This is why a few peoplewho are fully vaccinated do get sick from COVID19, though not usually asseriously as people who were not vaccinated. This doesn't mean that thevaccine "doesn't work", it just means that even trained immune systemscan sometimes be overwhelmed by a virus. <p> Because vaccinations train the immune system torecognize and fight off a virus, after a vaccination, <AHREF="https://www.mcgill.ca/oss/article/covid-19-health/i-felt-sick-after-getting-vaccine-why">wesometimes feel a bit sick</A>: some of the symptoms we experience when weare sick are caused by the body's defences: e.g. fever,aches, fatigue,and feeling unwell. In the case of a vaccination, though,this is not long-term, because a vaccination, unlike a virus, does notreproduce and spread, and so its effects will wear off quickly. <p> Vaccinations can sometimes cause side effects that are more serious. Thisis why they are tested carefully before approval. In the case of the majorCOVID19 vaccines, there are some very rare side effects that are serious: <AHREF="https://covid19-sciencetable.ca/sciencebrief/risk-of-vaccine-induced-thrombotic-thrombocytopenia-vitt-following-the-astrazeneca-covishield-adenovirus-vector-covid-19-vaccines/">certainCOVID19 vaccines cause very rare but quite serious blood clots</A>, and <AHREF="https://www.cdc.gov/coronavirus/2019-ncov/vaccines/safety/myocarditis.html">certainothers cause very rare heart inflammation</A>. These side-effects don't happen veryoften in people who receive the vaccine: they are much less likely than the<A HREF="https://www.nationalgeographic.com/science/article/flash-facts-about-lightning">probability of the average person being hit by lightning in their lifetime</A>. <p> The fact is, the vaccine is much less dangerous thanthe disease. A COVID19 infection can cause very serioushealth effects, and many of those effects are not rare.While most people who catch COVID19 recover at home, more than <AHREF="https://www.sciencedirect.com/science/article/pii/S0002934321000991">onein twenty</A> require hospitalization to stay alive. Of those, on the order of<A HREF="https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7920817/">onein ten die</A>. Moreover, many who recover from COVID19 <AHREF="https://www.mayoclinic.org/diseases-conditions/coronavirus/in-depth/coronavirus-long-term-effects/art-20490351">sufferlong-term health effects ranging from difficulty breathing, to fatigue, pain,and memory, concentration and sleep problems. Organ damage to the heart,lungs and brain is also possible. </A>COVID19 is spreading around the worldand most people will eventually be exposed to it. It is better to get thevaccine first, so that you are less likely to be harmed by the disease later. <p> There are claims on the Internet that COVID19 vaccines are muchmore dangerous than what I've written here. Many of these claims aremisunderstandings. Millions of people have received COVID19 vaccines.A few who have had health problems after receiving the vaccine have reportedtheir health problems as a possible "side effect" of the virus. In theUS, there is a <A HREF="https://vaers.hhs.gov">vaccine reporting systemcalled VAERS</A> where people can report bad health events that happened tothem after receiving a vaccine: this lets scientists investigate whetherthe vaccine might have caused the problem. If the vaccine is causinga particular health problem, that problem would happen more often topeople who receive the vaccine than to those who do not. But for mostof the health problems reported to VAERS, they are not happening moreoften to vaccinated people, they happen at roughly the same rate as theyhappen to anyone, and so the vaccine cannot be responsible. It appearsthat COVID19 vaccines cause very few serious health problems, and thoseare very rare. The evidence for this is that millions of people around theworld have received COVID19 vaccines and almost nobody has gotten seriouslysick from them. The COVID19 disease itself is much more dangerous, which iswhy hospitals are full of people suffering from the disease, not the vaccine. <p> Even so, wouldn't it be better to avoid both the vaccine and thedisease? Yes, it would be, if you could be assuredof never being exposed to COVID19. But that is not so easy. <AHREF="http://www.bccdc.ca/health-info/diseases-conditions/covid-19/about-covid-19/how-it-spreads">COVID19spreads very easily: it spreads through tiny moisture droplets in exhaledbreath</A> that float in the air like smoke from a cigarette, so if youare indoors with someone who is exhaling COVID19 virus, and there ispoor air circulation, you will inhale some. The longeryou are there, the more COVID19 virus you will inhale. <AHREF="https://www.nature.com/articles/d41586-021-02259-2">Not everyonewho gets COVID19 feels very sick right away: some feel fine, at leastfor a while</A>, and many who feel sick don't feel so sick that they stay home.They will spread the virus whereever they go, simply by exhaling. You may be in a room withan infected person who has no idea that they are spreading COVID19. This iswhy masks are so helpful, because the mask over the nose and mouth of aninfected person reduces the amount of COVID19 viruses they breathe out,and the mask over the nose and mouth of other people in the room reducesthe amount of COVID19 virus they might breathe in. It's also a reason whyindoor fresh air circulation is so important, and why COVID19 is so much moreof a danger indoors than outdoors. COVID19 is very contagious, especiallythe new "delta" variant which is the dominant variant circulating today: <AHREF="https://www.npr.org/sections/goatsandsoda/2021/08/11/1026190062/covid-delta-variant-transmission-cdc-chickenpox">onaverage, a sick person will spread it to six or more others. It's only alittle less transmissible than chickenpox, but a lot more transmissible thanflu.</A> It's quite possible that we will all be exposed to it eventually. <p>An even more important reason to be vaccinated is to reduce the spread ofCOVID19 to others. Remember that the only way for a virus to reproduceis in the body of an infected person. If most people make their bodiesinhospitable to the virus by getting vaccinated, then the virus will findvery few opportunities to spread. It's like fire trying to spread in avery wet forest: only the dry sticks will burn, and the fewer dry sticksthere are, the less likely the fire will find more sticks to spread to, andthe more likely it will burn out. So by getting vaccinated, we protect notonly ourselves, but everyone around us, especially those who, for medicalreasons, can't be vaccinated, or who have immune systems that don't workwell. If not enough of us get vaccinated, the number of COVID19 cases willoverwhelm the hospitals. Most of those who need hospital care for theirCOVID19 infections will die instead. Also, many people who need hospitalcare for other serious illnesses won't be able to get the care they need,and they will die too. <p>So please be brave: if you can, get vaccinated. Yes, the effects ofthe vaccine may be unpleasant for a few days as the body learns how tofight the virus. But the vaccine will not harm you like the disease will,and it will train the body's immune system to fight it. My father got COVID19 too early, lastsummer, before COVID19 vaccines were available. If they had been availablethen, he might still be alive today. They're available now. Please getvaccinated if you can. If enough people around the world get vaccinatedagainst COVID19, we may eventually be able to eliminate this diseasealtogether, and that would be a thing worth doing.</description> </item> <item> <title>The Covid19 Blues</title> <pubDate>Sun, 06 Jun 2021 13:39:00 -0400</pubDate> <link>http://www.cs.toronto.edu/%25252525252525252525257Ejdd/blog.cgi/2021/06/06#covidblues</link> <category>/misc</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%25252525252525252525257Ejdd/blog.cgi/misc/covidblues</guid> <description><figure style="float:left;width:40%;padding:1px"><IMG src="/~jdd/blog/misc/guitar-3291890_1280.jpg" alt="Man playing a guitar"><figcaption><small>Image by <a href="https://pixabay.com/users/lucasvieirabr-8401756/?utm_source=link-attribution&amp;utm_medium=referral&amp;utm_campaign=image&amp;utm_content=3291890">lucasvieirabr</a> from <a href="https://pixabay.com/?utm_source=link-attribution&amp;utm_medium=referral&amp;utm_campaign=image&amp;utm_content=3291890">Pixabay</a></small></figcaption></figure> <p>The arts find inspiration in times of trouble, none more so than the sort ofmusic known as the blues. Blues are creative and emotional,sometimes raw, but never fake. Blues are not about superstarsand megahits, blues are about the endurance and hope of ordinarypeople. As Covid19 drags on, endurance and hope are needed more thanever. Here are pointers to a few Covid19-inspired bluestracks that I appreciate. <ul><li> <A HREF="https://www.youtube.com/watch?v=MfNpa372t-Y">Keb' Mo'</A><li> <A HREF="https://www.youtube.com/watch?v=4bkPOgHRXPw">Rob Bentley</A><li> <A HREF="https://www.youtube.com/watch?v=1H31SiJC-8Q"> Mads Jacobsen</A><li> <A HREF="https://www.youtube.com/watch?v=hwHkYiPIZsk"> (Mads has an "Astronaut Remix" version too)</A><li> <A HREF="https://twitter.com/KarlKohlhase/status/1388477942192279556"> Karl Kohlhase</A></ul> Enjoy!</description> </item> <item> <title>What's Wrong With Passwords on the Internet Anyway?</title> <pubDate>Thu, 31 Dec 2020 22:57:00 -0500</pubDate> <link>http://www.cs.toronto.edu/%2525252525252525252525257Ejdd/blog.cgi/2020/12/31#passwords</link> <category>/it</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%2525252525252525252525257Ejdd/blog.cgi/it/passwords</guid> <description><figure style="float:left;width:40%;padding:1px"><IMG src="/~jdd/blog/it/registration-3938434_1280.jpg" alt="Completed Login Prompt"><figcaption><small>Image by <a href="https://pixabay.com/users/geralt-9301/?utm_source=link-attribution&amp;utm_medium=referral&amp;utm_campaign=image&amp;utm_content=3938434">Gerd Altmann</a> from <a href="https://pixabay.com/?utm_source=link-attribution&amp;utm_medium=referral&amp;utm_campaign=image&amp;utm_content=3938434">Pixabay</a></small></figcaption></figure> More than fifteen years ago, <A HREF="https://www.cnet.com/news/gates-predicts-death-of-the-password/">Bill Gates predicted that use of traditional passwords would dwindle.</a> This has happened to a certain extent, but a login and password is still the mostcommonly used credential for computing authentication. It isincreasingly problematic. According to <AHREF="https://enterprise.verizon.com/resources/reports/2020-data-breach-investigations-report.pdf">Verizon's 2020 Data Breach Investigations report, 37% of allbreaches involved the stealing of credentials or the use of stolencredentials. (p.7)</A> What is the root cause of the problem? <p> Put in simple terms, a login and password is what a system relies onto know who is who. Your password is secret: only you know what it is,and the system has some way of checking that it is correct. If someoneconnects to the system with your login and password, the system checksthat the password is the right one for your login. If it is, the systemconcludes that you are the person trying to connect, and lets you in.If you are the only one who knows the password, this approach works,since you are the only person who can provide the correct password. Butif criminals know your password too, and use it, the system will think thecriminals are you, and will give them access to your account and all yourdata. The only way to fix this is to change your password to somethingnew that only you know, but by then the damage may well be done. <p> Unfortunately, criminals have a pretty effective technique for findingout your login and password: they trick you into telling it to them. "Waita minute!", you might say, "I won't ever tell a criminal my password. I don'teven tell my family my password!" But you tell the system your passwordevery time you log in. So if criminals set up a fake system that lookslike the real one, and trick you into trying it, when you tell their fakesystem your password, the criminals will learn what it is. <p> This was not a common problem in the past, because it was difficultfor criminals to successfully set up fake systems that look convincing.But on the Internet today, it is easy to set up a web site that looks likeanother site. The only thing that's hard to fake is the first partof the link, the hostname section that comes immediately after the doubleslash (//) and before the first single slash (/), because that part of thelink is used to direct the request to the right system on the Internet. Butgiven that the Internet is available in hundreds of countries, each withits own set of internet service providers, it is often not too difficultfor criminals to find somewhere on the Internet where they can register ahostname that is similar-looking to the real thing. <p> Worse, the rise of messages containing embedded links make it very easyfor criminals to send a fake message (e.g. an email or text) with a link thatseems legitimate but really directs you to a fake site. This iscalled "phishing". Because of the way the web's markup language( HTML) works, it is easy to set up a link that seems to pointto one site, but actually points to another. For example, <AHREF="https://www.amazon.com">https://www.walmart.com</A> is a link thatseems to point to Walmart but really points to Amazon. Most web browserswill let you "hover" over a link to see where it really goes. But dopeople check every link carefully each time they use it? <p>The problem is made worse by the proliferation of legitimate messages with embedded links to all sorts of cloud services. I recently saw a messagefrom a large organization to its staff, about their pensions. The messagecontained links to an external site whose name had no resemblance tothe organization's name. The message invited the staff to click on thoselinks to see information about their pensions. The message was legitimate:the organization had contracted with an external cloud provider to providean online pension calculator for staff. But the message said nothingabout the cloud provider: it merely contained a link to the calculator.If criminals had sent a similar message containing a malicious link to afake system somewhere on the Internet, one that prompted staff to entertheir login and password, no doubt many staff would have thought itlegitimate. How could staff be expected to be able to tell the difference? <p>A good way to combat the password capturing problem is to requiremore than just a password to use a system. This is called "two-factor" or"multi-factor" authentication. Your password is one factor, and somethingelse is a second factor, and you must provide both factors to prove to thesystem that it is you. This helps because the criminals must have both yourpassword and your second factor in order to access your account and data. To ease the authentication burden for users, systems can ask for twofactors only sometimes, such as when logging in for the first time in awhile, or logging in from a new machine or a new location. <p> Ideally the second factor should be something that is hard forcriminals to capture and use. One problem with a password is that it is asecret that can be used from anywhere on the Internet. With <AHREF="https://www.statista.com/statistics/617136/digital-population-worldwide/">almost 60% of the world's population on the Internet</A>, which nowreaches every country in the world, the Internet can hardly be considered a"safe place". A second password, as easily used from anywhere on theInternet as the first, would not be much of an improvement. Worse would bethe answers to some personal question about yourself, such as your mother'smaiden name or the name of your first school: not only is such informationjust as easily used as a password, it is information that people may beable to find out in various ways. Answers to personal questions, whilesometimes used for authentication, typically do not make a good second factor. <p> A better second factor is a message sent via a communicationchannel that goes only to you: for example, an email to your emailaddress, or a text to your cell phone number. When you attempt tolog in, the system sends a unique one-time code to you through thatchannel, and asks you to enter it. The assumption is that criminalswon't have access to your email or your cell number, so they won't knowand be able to enter the one-time code that the system sent to you. Thisis usually a good assumption. But criminals can try to get accessto your email or your phone number, and sometimes they succeed. Forexample, in the case of a cell number, one thing they could try is to <AHREF="https://www.howtogeek.com/358352/criminals-can-steal-your-phone-number.-heres-how-to-stop-them/">callyour cell phone provider, tell them they are you and that your phone hasbeen stolen, and request that your phone number be transferred to theirnew phone.</A> <p> Another second factor, one even better, is a physical device in yourpossession. This could be a hardware security token that you plug intoyour computer or that displays a unique, frequently changing, code. Or itcould be an app on your cell phone that is tied to your unique device. Aphysical device is an excellent second factor, because most criminals onthe Internet are physically distant. To successfully pretend to be you,a criminal would need direct physical access to a device that would likelybe located in your purse or pocket. <p> Relying on a device in purse or pocket as well as a password in your headis an improvement in security, but it has its drawbacks. It makes that deviceessential for you to use the system: if it is broken, lost or stolen,you're locked out, even if you know the password. While locking out peoplewho don't have the device is exactly the point, that doesn't help when it iskeeping you from legitimately using the system. Moreover, if that deviceis your smartphone, it changes your phone from a convenience to anecessity. While a smartphone has become a necessity already to some,it is a potentially consequential thing for it to become a requirement foreveryone. A hybrid approach is perhaps best: hardware security tokens thosewho prefer it, a smartphone for those who for their own reasons carry onearound anyway, and for many, both: a smartphone for convenience, with ahardware security token as backup, in case of smartphone loss or damage. <p> Perhaps there is an even more secure option? What if your secondfactor wasn't a device, but an actual physical part of your body,such as a finger (for a fingerprint), eye (for a retinal scan), face,or even heartbeat (as measured by e.g. a <A HREF="https://www.nymi.com/">Nymi Band</A>)? Would that be better still? After all, if it is hardfor a criminal to get access to someone's things without being noticed,it is even harder to get access to someone's body. This is indeedpossible: a technique called "biometrics, and it can bean effective second factor. Unfortunately there are a couple of issueswith biometrics. For example, injuries or health issues can change yourbody; a cut on your finger may affect your fingerprint, for instance.Secondly, biometrics have a "revocation" problem. This comes from the factthat a biometric is a unique measurement of your body part: a fingerprint, retinal scan, facial image, or ECG. But measurementsare data, and biometric data, like any other data, can and <AHREF="https://www.forbes.com/sites/zakdoffman/2019/08/14/new-data-breach-has-exposed-millions-of-fingerprint-and-facial-recognition-records-report/">has been breached.</A> If this happens, what will you do? Passwords canbe changed, hardware security tokens can be replaced, but how are yougoing to change your fingerprint, your face, your eye, your heartbeat?While biometrics do have a place in authentication, most commonly to unlocka local device such as a smartphone or a laptop, the lack of revocabilitymake biometrics less suitable as a second factor for Internet-accessible services. <p> Regardless of what is chosen for a second factor, the inconvenienceof using more than one factor is something that has to be considered.Passwords, especially ones that are easy to remember, arequite convenient. Requiring more than this can make authentication moredifficult. If becomes too difficult, the difficulty becomes a disincentiveto use the system. For systems protecting highly sensitive data, somedifficulty may be warranted, given the risk. For lower-risk systems, thingsare less clear. Yet for Internet-accessible systems, due to the prevalenceof phishing, something more secure than just passwords seems increasinglynecessary. I think Bill Gates is right: like it or not, the traditionalpassword will become increasingly rare on the Internet, for good reason.</description> </item> <item> <title>Thoughts on Covid19</title> <pubDate>Mon, 23 Nov 2020 00:00:00 -0500</pubDate> <link>http://www.cs.toronto.edu/%252525252525252525252525257Ejdd/blog.cgi/2020/11/23#covid</link> <category>/misc</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%252525252525252525252525257Ejdd/blog.cgi/misc/covid</guid> <description><figure style="float:left;width:40%;padding:1px"><IMG src="/~jdd/blog/misc/virus-4937553_1280.jpg" alt="Visual representation of Covid19 viruses under electron microscope"><figcaption><small>Image by <a href="https://pixabay.com/users/PIRO4D-2707530/?utm_source=link-attribution&amp;utm_medium=referral&amp;utm_campaign=image&amp;utm_content=4937553">PIRO4D</a> from <a href="https://pixabay.com/?utm_source=link-attribution&amp;utm_medium=referral&amp;utm_campaign=image&amp;utm_content=4937553">Pixabay</a></small></figcaption></figure> I'd recently reread a blog entry I'd written more than a year ago nowon intentionality about blog posting. After writing it, I lived it: I wroteseveral additional blog entries throughout the year. But then along camethe Covid19 pandemic, and it illustrated a problem with intentionality:intentionality requires priority. When Covid19 hit Ontario in March, the pandemicrequired substantial changes in how I live and work, and that drove areprioritization of my efforts, both in my job as Director responsiblefor computing at the University of Toronto's Computer Science department,and at home, as a parent of teenagers in high school. In the face ofthe challenges of Covid19, blogging seemed not sufficiently important,and of course, it wasn't. So I didn't write, I worked. I am grateful tohave work, in fact: I know of others who couldn't work because the sortof work they did couldn't be done from home. I consider myself fortunateto work in the computing field, which has not been so badly affected. Infact, in many ways, computing has been part of the solution (networking,videoconferencing, cloud computing, medical informatics, etc.) and has beenboosted rather than impaired. In my job, I and my staff, and my department,found ourselves not without work, but with too much. This is not necessarilya bad situation to be in, but it doesn't lend itself to blogging. <p> Another reason is that Covid19 didn't just affect me professionally, itaffected me personally: I lost a parent to Covid19 this summer.While I am not in any way unique in having lost someone to this disease,I was not really in a good state to blog, for quite some time. <p> There is still another factor, though, one that also kept me fromblogging. I am no epidemiologist. Still, as a thinking person, I seek tounderstand what was going on, why, and what can be done about it.Seeking to understand is, for me, theraputic: it helps me deal with stress,anxiety, grief, and loss. <p> First, I looked for good sources of information about thepandemic itself. The <A HREF="https://www.cdc.gov"> Centrefor Disease Control and Prevention</A> in the US has plenty ofgood material about it. One thing I found particularly helpful was an <AHREF="https://www.cdc.gov/mmwr/volumes/69/wr/mm6919e6.htm">analysis inmid-May</A> about a choir practice in Washington state with 61 attendees, onethat led to most becoming infected. It resulted in three hospitalizations andtwo deaths. The CDC report is a very helpful example of rigorous statisticaldata analysis set in a small, understandable real-world context. As anillustration of what the Covid19 virus is like, I find it very helpful. Forinstance, it suggested airborne spread before that became generally realized. <p> Secondly, information about previous pandemics. Again,the <A HREF="https://www.cdc.gov">Centre forDisease Control and Prevention</A> in the US has a <AHREF="https://www.cdc.gov/flu/pandemic-resources/basics/past-pandemics.html">verygood past pandemics page</A>, put together before the Covid19 pandemicstarted, covering the horrifying 1918 influenza pandemic that killed fiftymillion people around the world, and the later influenza epidemics of 1957,1968, and 2009. Each of these provide a general helpful picture: firstly,that each pandemic has a timeframe that is typically greater than one yearbut less than two, that transmission reduces in the summer but increasesin the fall/winter due to indoor crowding and decreased relative humidity,and that mass vaccination can be an effective way to ward off a disasterof the scale of the 1918 pandemic. <p> One problem with this current pandemic is that, unlike the pandemics of1957, 68, and 2009, the virus is not influenza, but a coronavirus. Thereare four coronaviruses that have been circulating widely for years <AHREF="https://www.cdc.gov/coronavirus/types.html">(229E, NL63, OC43, andHKU1</A>), but they typically don't cause serious illness. Two others(SARS-CoV and MERS-CoV) emerged in the early 21st century, both quitedangerous and certainly serious enough to warrant vaccination were theyto spread widely, but due to a great deal of diligence and effort, and nota little good fortune, both of these were kept from spreading throughthe world population. The current Covid19 pandemic,caused by yet another coronavirus, SARS-CoV2, is the first coronavirusboth serious enough and widespread enough to warrant a vaccine. Unlikefor influenza, a coronavirus vaccine has never been produced before, so it hastaken longer than it would if this pandemic had been influenza. Only now,as we approach the one year mark of the virus' first emergence, are weseeing some likely vaccine candidates. It will still take some time toproduce and distribute suitable vaccines. <p> In the meantime, while efforts continue to design,test, produce and distribute a suitable vaccine,the challenge is to keep Covid19 from spreading far andfast. While at first it was believed that <AHREF="https://www.canada.ca/en/public-health/services/diseases/2019-novel-coronavirus-infection/health-professionals/main-modes-transmission.html">Covid19spreads primarily through surface contact, there is increasing evidencefor areosol spread (fine droplets in the air)</A>. So methods are neededto hinder the passing of the virus from one person to another. Thereare two main approaches: keeping people further apart, and putting physicalbarriers (e.g. masks) and processes (e.g. handwashing) in place so thatthe virus can't easily pass from one person to another.<p>The best way to hinder the transmission of Covid19 is to find out who maybe contagious (through testing and contact-tracing), and keepthem away from everyone else (quarantine) until they are no longer contagious.One challenge is that it can sometimes be very hard to detectwhen someone has Covid19 and is spreading the virus. Thereis a wide variation in how Covid19 affects people who have it. Formany, it can take days for symptoms to emerge (presymptomatic), and <AHREF="https://www.healthline.com/health-news/20-percent-of-people-with-covid-19-are-asymptomatic-but-can-spread-the-disease">forsome, Covid19 can be mostly or completely asymptomatic, yet asymptomatic andpresymptomatic Covid19 patients can spread the disease</A>. If those whomay have Covid19 can be identified (through testing and thorough contacttracing), then those individuals alone can be quarantined until they areno longer contagious. If they cannot be identified, then the only wayto hinder the spread of the disease is to assume that almost anyone mighthave Covid19. This requires such things as requiring everyone to wear masks,and, despite severe social and economic cost, lockdowns, which are a sort ofsemi-quarantine for everyone. As I write this, Covid19 has been spreadingquite quickly in my city, Toronto, despite a mask mandate, and so <AHREF="https://www.toronto.ca/news/city-of-toronto-supports-provincial-lockdown-measures-to-reduce-covid-19-transmission/">Torontois going into lockdown</a>. <p> How will it all end? In the struggle betweenpessimism and hope, I choose hope. I hope that I will not lose any morefamily members to this disease. I hope that effective vaccines will soonbe available in the necessary quantities. I hope that the measures takento hinder the spread will be effective. I think it is reasonable to expectthat we will see the widespread distribution of effective vaccines in 2021,and this pandemic will be over sometime next year. Will everything be thesame? No, I think not. Some businesses (tourism and travel, for example)will have a massive economic hole to climb out of, and some companies willnot survive, but people will travel again. Working from home, and technologyin support of it, will be more widely accepted. Cheek-to-jowl "open-concept"offices, handshaking, and other close-quarters working practices willbe less readily accepted. There will be a greater consciousness of viralhygiene, and a greater acceptance of masks. But life will go on. Covid19will no longer command the attention it is getting now. Other things willseem important again. And there will be many worthwhile things to blog about.</description> </item> <item> <title>Some Clarity on Public Cloud Cybersecurity</title> <pubDate>Mon, 24 Feb 2020 10:19:00 -0500</pubDate> <link>http://www.cs.toronto.edu/%25252525252525252525252525257Ejdd/blog.cgi/2020/02/24#cloudsecurity</link> <category>/it</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%25252525252525252525252525257Ejdd/blog.cgi/it/cloudsecurity</guid> <description><figure style="float:left;width:40%;padding:1px"><IMG src="/~jdd/blog/it/clear-114143_640.jpg" alt="Break in clouds, revealing clear skies"><figcaption><small>Image by <a href="https://pixabay.com/users/36882-36882/?utm_source=link-attribution&amp;utm_medium=referral&amp;utm_campaign=image&amp;utm_content=114143">Sabrina Corana</a> from <a href="https://pixabay.com/?utm_source=link-attribution&amp;utm_medium=referral&amp;utm_campaign=image&amp;utm_content=114143">Pixabay</a></small></figcaption></figure> I've been thinking about public cloud cybersecurity for some years now,as I've watched adoption of the public cloud grow from a trickle to aflood. Early on, most of the reasons I heard for public cloud adoptionmade a great deal of sense to me: the need to rapidly scale up anddown the size of a service, the desire to leverage the expertiseof a large technical partner with resources in network and computinginfrastructure exceeding one's own, the desire to leverage geographicallydiverse, redundant datacentres, the desire to fund computing fromoperating rather than capital budgets, and the desire to build adaptable,scriptable services with better connectivity to the Internetthan one could otherwise provide for oneself. But in the last yearor two, as anxiety about cybersecurity increases, I've been hearingmore and more people refer to cybersecurity as their primary reasonfor their adoption of the public cloud. I'm not so sure what I thinkof this reasoning. I can understand why someone might want to passto a third party a task that makes them anxious. Insituations involving strong emotions, such as anxiety, there is risk of <AHREF="https://www.psychologytoday.com/ca/blog/science-choice/201504/what-is-confirmation-bias">"confirmation bias": believing something is true because you want it to betrue.</A> But is it? <AHREF="https://www.investopedia.com/terms/c/ceterisparibus.asp">Ceterisparibus (all other things being equal)</A>, is the public cloud intrinsiclymore secure than on-premise datacentres? <P>Some argue yes. Eplexity calls cloud computing <AHREF="https://eplexity.com/why-the-public-cloud-is-more-secure-than-an-on-premises-data-center/">"anestablished best practice for businesses" and claims"your data is typically safer in the public cloud than in anon-premises data centre".</A> In 2016, SaraPatrick of Clutch, guest-writing for Tripwire.com, claimed to have <AHREF="https://www.tripwire.com/state-of-security/security-data-protection/4-reasons-why-the-cloud-is-more-secure-than-legacy-systems/">"fourreasons why the Cloud is more secure than Legacy Systems"</A> In 2017,Quentin Hardy of the New York Times claimed that cloud data is <AHREF="https://www.nytimes.com/2017/01/23/insider/where-does-cloud-storage-really-reside-and-is-it-secure.html">"probably more secure than conventionally stored data."</A>In 2018, David Linthicum, writing for InfoWorld, claimed <AHREF="https://www.infoworld.com/article/3316637/how-the-cloud-has-made-you-more-secure.html">"your information is actually safer in the cloud than it is in your own data centre"</A>. <P> One reason given for the claim is that public cloud providers offer greatertechnical expertise than what is possible on-premise. Eplexity writes:<A HREF="https://eplexity.com/why-the-public-cloud-is-more-secure-than-an-on-premises-data-center/"><blockquote> Unless your company is already in the business of IT security, spending time and effort on securing your on-premises data distracts from your core functions. Most organizations likely don't have a robust, experienced team of cybersecurity professionals at their disposal to properly protect their on-premises data. ... As such, cloud providers may employ hundreds or thousands of developers and IT professionals.</blockquote> </A> This is an argument from size and scale. Cloud providers are bigger than you,and have arguably more IT expertise than you, so they can do a better jobthan you. But sadly, size and IT expertise is no guarantee of security. <AHREF="https://qz.com/741056/the-stunning-collapse-of-yahoos-valuation/">Yahoowas a large Internet company, valued at one time at $125 billion.</A>It employed thousands of developers and IT professionals. Yet it was <AHREF="https://www.nytimes.com/2017/10/03/technology/yahoo-hack-3-billion-users.html">subjectto a cybersecurity breach of three billion user accounts in 2013/14; thebreach was not disclosed until the fall of 2016, and the full impact wasnot known until october 2017. The damage to Yahoo's business was significant:Verizon acquired Yahoo in 2017 for less than $5 billion, a deal that wasnearly derailed by the disclosure of the breaches.</A> <P>I think we must conclude from the Yahoo story that size and expertisealone is no guarantee of cybersecurity. Naturally, major cloud providerslike Amazon, Microsoft and Google are aware of the Yahoo situation andits consequences. No doubt it illustrated for them the negative impactthat a major breach would have on their business. I cannot imagine thatthey would take the threat lightly. <P>Yet there have been close calls. Microsoft, a major cloud provider, <AHREF="https://www.zdnet.com/article/microsoft-discloses-security-breach-of-customer-support-database/">in December 2019 accidentally disclosed to the world a cloud databaseon Azure with 250 million entries of customer support data. Happily,a security researcher spotted and reported it, and Microsoft fixed it soon after.</A> Moreover, <AHREF="https://www.forbes.com/sites/zakdoffman/2020/01/30/severe-perfect-100-microsoft-flaw-confirmed-this-is-a-cloud-security-nightmare/#7b5493dfb4a4">ZakDoffman, writing for Forbes, reported in Jan 2020 that Check Point SoftwareTechnologies, a cybersecurity vendor, had discovered in 2019 a serious flawin Microsoft Azure's infrastructure that allowed users of the service toaccess other users' data.</A> While Check Point reported it immediately toMicrosoft, who fixed it quickly, had the flaw been discovered by criminalsinstead of cybersecurity researchers, a great many things running on Azurecould have been compromised. Doffman quotes Yaniv Balmas of Check Point: <AHREF="https://www.forbes.com/sites/zakdoffman/2020/01/30/severe-perfect-100-microsoft-flaw-confirmed-this-is-a-cloud-security-nightmare/#7b5493dfb4a4"><blockquote> ...the take away here is that the big cloud concept of security free from vulnerabilities is wrong. That's what we showed. It can happen there as well. It's just software and software has bugs. The fact I can then control the infrastructure gives me unlimited power.</blockquote></A> In the Check Point research article describing the flaw, Balmas concludes: <A HREF="https://research.checkpoint.com/2020/remote-cloud-execution-critical-vulnerabilities-in-azure-cloud-infrastructure-part-ii/"><blockquote>The cloud is not a magical place. Although it is considered safe, itis ultimately an infrastructure that consists of code that can havevulnerabilities - just as we demonstrated in this article.</blockquote></A> <P>What, then, is the right answer? Well, there isn't one. Neither publiccloud or on-premise datacentres are magic, neither are "safe". Cybersecurityis a challenge that has to be met, no matter where the service is, or whatinfrastructure it is using. Happily, this is finally beingrecognized. Even Gartner Research, a long-time proponentof the public cloud, predictingas recently as mid-2019 that <AHREF="https://web.archive.org/web/20190808014445/https://www.gartner.com/smarterwithgartner/is-the-cloud-secure/">publiccloud infrastructure as a service (IaaS) workloads will suffer at least60% fewer security incidents than those in traditional data centers</A>, has recently taken a more nuanced view.In the fall of 2019, this prediction of fewer security incidents in the cloud disappeared from Gartner's website,and was replaced by this:<A HREF="https://web.archive.org/web/20191130104612/www.gartner.com/smarterwithgartner/is-the-cloud-secure/"><blockquote> <B>Through 2024, the majority ofenterprises will continue to struggle with appropriately measuring cloudsecurity risks.</B> <BR>Questions around the security of public cloudservices are valid, but overestimating cloud risks can result in missedopportunities. Yet, while enterprises tended to overestimate cloud riskin the past, there's been a recent shift - many organizations are nowunderestimating cloud risks. This can prove just as detrimental, if notmore so, than an overestimation of risk. A well-designed risk managementstrategy, aligned with the overarching cloud strategy, can help organizationsdetermine where public cloud use makes sense and what actions can be takento reduce risk exposure. </blockquote> </A><P> So does "public cloud use make sense"? Yes, of course it does, for a greatmany things. But it's not because the public cloud is intrinsicly more secure. Thepublic cloud has its own set of cybersecurity issues. There is no "free pass".As always, carefully assess your risks and make an informed decision.</description> </item> <item> <title>Does AI Help or Hinder Cybersecurity?</title> <pubDate>Fri, 24 Jan 2020 20:02:00 -0500</pubDate> <link>http://www.cs.toronto.edu/%2525252525252525252525252525257Ejdd/blog.cgi/2020/01/24#aisecurity</link> <category>/it</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%2525252525252525252525252525257Ejdd/blog.cgi/it/aisecurity</guid> <description><figure style="float:left;width:50%;padding:1px"><IMG src="/~jdd/blog/it/security-4700815_640.jpg" alt="Hooded figure with glowing circuit-board visage"><figcaption>Image by <a href="https://pixabay.com/users/geralt-9301/?utm_source=link-attribution&amp;utm_medium=referral&amp;utm_campaign=image&amp;utm_content=4700815">Gerd Altmann</a> from <a href="https://pixabay.com/?utm_source=link-attribution&amp;utm_medium=referral&amp;utm_campaign=image&amp;utm_content=4700815">Pixabay</a></figcaption></figure> Both AI and cybersecurity have become increasingly prominent in recentyears. AI's prominence has been driven by advances in machine learning andthe very real improvements it has made in the ability of computersystems to do things that previously seemed possible only to humanbeings. Cybersecurity's prominence has been driven by a number ofdevelopments, including increasing nation-state conflict on the Internet,and a dramatic rise in organized cyber-crime. It is inevitable that the twowill combine: AI will be and is being applied to the cybersecurity space,through the development of machine learning techniques for breaking intoand defending systems. <P> One view on this is that machine learning, as a powerful techniquethat enables computer systems to take on tasks previously reserved onlyfor humans, will empower cyberattackers to breach computer security innew ways, or at least in ways more effective than before. I know there isa great deal of anxiety about this. This past fall, I had a conversation witha CIO of a large university, who told me that his university was migratingits internet services to Amazon precisely because he believed that newAI-powered cyberattacks were coming, and he thought Amazon would be betterable to fend them off. I'm not sure what I think of this defensive strategy,but that is not the important question here. The key question is this:are AI-powered cyberattacks going to overwhelm cyberdefence? <P> No doubt AI-powered cyberattacks are a reality. Machine learning is apowerful computer science technique, especially for automation. Cyberattackers, especially sophisticated, well-funded cyberattackers,will use it and I am confident are already using it. But highlyautomated cyberattacks are nothing new: cyberattackers have been automatingtheir attacks for decades. Smarter automated cyberattacks are certainlysomething to worry about, but will they be transformative? Maybe. Afterall, in cybersecurity, the advantage is to the attacker, who needs to findonly one hole in the defences, while the defender needs to block all of them.Anything that boosts the effectiveness of the attacker would seem tomake the situation worse. <P> To really see the full picture, it's important to look at thedefender too. Machine learning makes the situation worse only if itbenefits the attacker more than it benefits the defender. But does it? <P> I don't have a complete answer to this question: there is a greatdeal of work still to be done on the application of machine learningto cybersecurity. But I suspect that the answer is a qualified <em>No</em>:rather, all other things being equal, machine learning will likely shiftthe balance of power towards the defender. The reason is data. <P> Machine learning is a technique where computer systems, instead of beingprogrammed by programmers, learn what to do from data. But the quality ofthe learning depends on the quality and in particular the quantity of data.Machine learning is a technique that is most effective when trained withlarge amounts of data. <A HREF="http://www.image-net.org">ImageNet</A>,for instance, a standard training dataset used to train machine learningapplications to recognize images, contains about 14.2 million images. But who is more likely to have access to large amounts of good dataabout a system: the attacker or the defender? Of course, it depends, butit seems to me that, very generally speaking, the defender is more likelyto have access to good system data than the attacker. The attacker istrying to get in; the defender is already in. <P> Of course, this is the broadest of generalizations. The effectivenessof machine learning in the cybersecurity space depends on a great manythings. But I am cautiously optimistic. I realize I may be bucking whatseems to be becoming a prevailing trend of ever-increasing anxiety aboutcybersecurity, but I believe here that machine learning has more potentialto help than to harm. I look forward to seeing what will emerge in thisspace over the next few years.</description> </item> <item> <title>What's all the fuss about AI anyway?</title> <pubDate>Mon, 30 Sep 2019 00:00:00 -0400</pubDate> <link>http://www.cs.toronto.edu/%252525252525252525252525252525257Ejdd/blog.cgi/2019/09/30#ai</link> <category>/it</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%252525252525252525252525252525257Ejdd/blog.cgi/it/ai</guid> <description><figure style="float:left;width:50%;padding:1px"><IMG src="/~jdd/blog/it/images/a-2729794_640.png" alt="Brain-shaped Network"><figcaption>Image by <a href="https://pixabay.com/users/GDJ-1086657/?utm_source=link-attribution&amp;utm_medium=referral&amp;utm_campaign=image&amp;utm_content=2729794">Gordon Johnson</a> from <a href="https://pixabay.com/?utm_source=link-attribution&amp;utm_medium=referral&amp;utm_campaign=image&amp;utm_content=2729794">Pixabay</a></figcaption></figure> A great deal in the past five years has been written about ArtificialIntelligence (AI). But there's a lot of confusion about what AI actuallyis, and why it is of special interest now. Lets clear up some of thatconfusion. In ordinary language, what is this fuss about AI all about? <P>AI, broadly understood, is a term used to describe a set of computingtechniques that allow computers to do things that human beings use intelligence to do. This is not to say that the computer is intelligent, butrather that the computer is doing something that, if done by a person, wouldbe considered evidence of that person's intelligence. Contrary to widespreadopinion, this is not the same thing as an artificial person. In fact, therehave been for a long time many things that humans use intelligence to do,that computers do better, whether it be remembering andrecalling items, doing arithmetic, or playing chess. But computersdo these things using different techniques than humans do. For example, <AHREF="https://www.chess.com/article/view/deep-blue-kasparov-chess">DeepBlue, a custom chess computer built by IBM, beat Garry Kasparov, thethen-reigning world chess champion, in 1997</A>, but Deep Blue played chessin a very different way than Garry. Garry relied on his human intelligence,while Deep Blue used programming and data. <P>However, some computer scientists, noting that people can do things thatcomputers can't, thought long and hard about ways that people do it,and how computers might be progammed to do the same. One such technique,deep learning, a neural network technique modelled after the human brain,has been worked on since the 1980s, with slow but steady improvement,but computer power was limited and error rates were often high, and formany years, most computer scientists seemed to feel that other techniqueswould yield better results. But a few kept at it, knowing that the computersof the day were inadequate, but advances in computing would make thingspossible that weren't possible before. <P>This all changed in 2012, when one such researcher,Geoff Hinton, and his students, working here at the University of Toronto, published a <AHREF="https://papers.nips.cc/paper/4824-imagenet-classification-with-deep-convolutional-neural-networks.pdf">seminaldeep learning paper that cut error rates dramatically.</A>I remember supporting Geoff's group's research computing at that time.It was a bit challenging: we were using multiple GPUs per machine totrain machine learning models at a time when GPU computing was still rathernew and somewhat unreliable. But GPUs were absolutely necessary: withoutthem, instead of days of computing time to train a model, months would berequired. One of our staff, Relu Patrascu, a computer scientist and skilledsystem administrator working hand-in-glove with the researchers, tuned andconfigured and babysat those machines as if they were sick children. But itworked! Suddenly deep learning could produce results closer to what peoplecould do, and that was only the beginning. Since then, deep learning hasproduced terrific results in all sorts of domains, some exceeding whatpeople can do, and we've not even scraped the surface of what is possible. <P> But what does deep learning actually do? It is a computerscience data classification technique. It's used to take input data andclassify it: give it a thing and it will figure out what the thing is.But it classifies things in a way that's different and more useful thantraditional computer science methods for classification, such as computerprogramming, or data storage and retrieval (databases). As such, it can beused to do a lot more than computers previously had been able to do. <P>To see this, consider traditional computer science methods: for example,computer programming. This approach requires a person to write code thatexplicitly considers different cases. For example, imagine that you want toclassify two-dimensional figures. You want to consider whether they areregular polygons. You could write a computer program that defines for itselfwhat a regular polygon is, and checks each characteristic of an input shapeto see whether or not it matches the definition of a regular polygon. Such aprogram, when given a square, will notice that it is a polygon, it has foursides, and that those sides are equal in length. Since the programmerput into the program a detailed definition of what a regular polygon is,and since the program checks each feature explicitly, it can tell whetheror not a shape is a regular polygon, even if the program has never seenthat particular shape before. <P>But what about exceptional cases? Is a circle a regular polygon? It is,after all, the limit of an N-gon as N goes to infinity. This is an "edgecase" and programs need to consider those explicitly. A programmer had toanticipate this case and write it into the program. Moreover, if you wanted toconsider some other type of shape, a programmer would have to rewritethe code accordingly. There's no going from a bunch of examples to workingcode without a programmer to write it. Programming is certainly a usefultechnique, but it has its limits. Wouldn't it be nice to be able tolearn from a bunch of examples, without a person having to write all thatcode? <P> One way to do that would be data storage and retrieval, for example,a database. Consider the shape classifier problem again. You might putin a bunch of shapes into a database, indicating whether the shape isa regular polygon or not. Once the database is populated, classifying ashape simply becomes looking it up. The database will say whether or notit is a regular polygon. <P> But what if it's not there? A database has the advantage of beingable to learn from examples. But it has a big disadvantage: if it hasn'tseen an example before, and is asked about it, it has no idea what theright answer is. So while data storage and retrieval is a very usefulcomputing technique, and it is the backbone of most of our moderninformation systems, it has its limits. Wouldn't it be nice ifa classifier system could provide a useful answer for input datathat it's never seen before, without a programmer to tell it how? <P> Deep learning does exactly this. Like data storage and retrieval, itlearns from examples, through training. Very roughly, a neural network,when trained, is given some input data, and is told what outputdata it should produce when it sees that data in future. These input andoutput constraints propagate forward and backwards through the network,and are used to modify internal values such that when the network nextsees input like that, it will produce the matching output. <P> The key advantage of this technique is that if it sees data thatis similar to, but not the same as data it has been trained on, it willproduce output similar to the trained output. This is very important,because like programming, it can work on input it has never seen, but likedatabases, it can learn from examples and need not be coded by a programmeranticipating all the details in advance. For our shape example, if trainedwith many examples of regular polygons, the neural network will be able tofigure out whether or not a given input is a regular polygon, and perhapseven more interestingly, it will be able to note that a circle is verylike a regular polygon, even if it had never been trained on a circle. <P> Moreover, a deep learning neural network can learn from its ownresults. This is called reinforcement learning. This technique involvesusing a neural network to derive output data from some input data, theresults are tested to see how well they work, and the neural network isretrained accordingly. This way a neural network can "learn from its ownmistakes", training itself iteratively to classify better. For example, amodel of a walking human, with some simple programming to teach it the lawsof physics, can, using reinforcement learning, teach itself how to walk. Afew years ago, some of the researchers in our departmentdid exactly that. Another example: Google got a lot ofattention a few years ago when deep learning researchers there <AHREF="https://deepmind.com/research/case-studies/alphago-the-story-so-far">built a deep learning system that used reinforcement learning to become achampion at the game of Go, a game very hard to computerize using traditionaltechniques, and proved it by beating the reigning Go world champion. </A> <P> It seems clear to me at this point that deep learning is as fundamentala computing technique as computer programming and databases in buildingpractical computer systems. It is enormously powerful, and is causing agreat deal of legitimate excitement. Like all computer science techniques, ithas its advantages and drawbacks, but its strengths are where other computerscience techniques have weaknesses, and so it is changing computer science(and data science more generally) in dramatic ways. It's an interesting timeto be a computer scientist, and I can't even begin to imagine the many thingsthat bright and innovative people will be able to do with it in the future.</description> </item> <item> <title>Existential threats from AI?</title> <pubDate>Mon, 02 Sep 2019 20:14:00 -0400</pubDate> <link>http://www.cs.toronto.edu/%25252525252525252525252525252525257Ejdd/blog.cgi/2019/09/02#aithreats</link> <category>/it</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%25252525252525252525252525252525257Ejdd/blog.cgi/it/aithreats</guid> <description><figure style="float:left;padding:1px"><IMG src="/~jdd/blog/it/nuclear-2136244_640.jpg" alt="Nuclear explosion"><figcaption>Image by <a href="https://pixabay.com/users/AlexAntropov86-2691829/?utm_source=link-attribution&amp;utm_medium=referral&amp;utm_campaign=image&amp;utm_content=2136244">Alexander Antropov</a> from <a href="https://pixabay.com/?utm_source=link-attribution&amp;utm_medium=referral&amp;utm_campaign=image&amp;utm_content=2136244">Pixabay</a></figcaption></figure> Plenty has been written about the possible threats to humanity from ArtificialIntelligence (AI). This is an old concern, a staple of science fiction sinceat least the 1950s. The usual story: a machine achievessentience and pursues its own agenda, harmful to people. The current successesof machine learning have revived this idea. The <A HREF="https://www.bbc.com/news/technology-30290540">late StephenHawking warned the BBC in 2014</A> that "the development of full artificialintelligence could spell the end of the human race". He feared that "itwould take off on its own, and re-design itself at an ever increasing rate."He worries that human beings, "who are limited by slow biological evolution, couldn't compete, and would be superseded."<AHREF="https://www.theatlantic.com/magazine/archive/2018/06/henry-kissinger-ai-could-mean-the-end-of-human-history/559124/">HenryKissinger, in a thoughtful essay in <em>The Atlantic</em> lastyear</A>, worried that "AI, by mastering certain competencies morerapidly and definitively than humans, could over time diminish humancompetence and the human condition itself as it turns it into data." <AHREF="https://www.wired.com/story/elon-musk-humanity-biological-boot-loader-ai/">ElonMusk, in a debate last month with Alibaba's Jack Ma, reported by WIRED,</A> argued that "there's just a smaller and smaller corner ofwhat of intellectual pursuits that humans are better than computers. Andthat every year, it gets smaller and smaller, and soon will be far farsurpassed in every single way. Guaranteed. Or civilization will end." <P>Are they right? Is there an existential threat to humanity fromAI? Well, yes, I think there actually is one, but not quite in the wayMusk, Kissinger, or Hawking fear. Computer have been better at humans fora long time in many cognitive domains. Computers rememberthings more accurately, process things faster, and scale betterthan humans in many tasks. AI, particularly machine learning, increasesthe number of skills where computers are better than humans. Given thathumanity has been spending the last couple of generations getting used toa certain arrangement where computers are good at some things and humansare good at others, it can be a bit disconcerting to have this upended bycomputers suddenly getting good at things they weren't good at before. Iunderstand how this can make some people feel insecure, especially highlyaccomplished people who define themselves by their skills. Kissinger,Musk and Hawking fear a world in which computers are better at many thingsthan humans. But we have been living in such a world for decades. AI simplybroadens the set of skills in question. <P>As a computer scientist, I am not particularly worried about the notion ofcomputers replacing people. Yes, computers are developing new useful skills,and it will take some getting used to. But I see no imminent danger ofAI resulting in an artificial person, and even if it did, I don't thinkan artificial person is an intrinsic danger to humans. Yet I agree thatthere are real existential threats to humanity posed by AI. But these arenot so much long term or philosophical, to me they're eminently practicaland immediate. <P> The first threat is the same sort of threat as posed by nuclearphysics: AI can be used to create weapons that can cause harm to people on a massive scale. Unlike nuclear bombs, AI weapons do not do their harm throughsheer energy discharge. Rather, machine learning, coupled with advances inminiaturization and mass production, can be used to create horrific smartweapons that learn, swarms of lethal adaptive drones that seek out anddestroy people relentlessly. A deep commitment to social responsibility,plus a healthy respect for the implications of such weapons, will be neededto offset this danger. <P> The second threat, perhaps even more serious, comes not from AI itselfbut from the perceptions it creates. AI's successes are transforminghuman work: because of machine learning, more and more jobs,even white-collar ones requiring substantial training, can be replaced bycomputers. It's unclear yet to what extent jobs eliminated by AI will be offsetby new jobs created by AI, but if AI results in a widespread perceptionthat most human workers are no longer needed, this perception may itselfbecome an existential threat to humanity. The increasingly obvious factof anthropogenic climate change has already fueled the idea that humanityitself can be viewed as an existential threat to the planet. If AI makesit possible for some to think that they can have the benefits of societywithout keeping many people around to do the work, I worry we may seeserious consideration of ways to reduce the human population to muchsmaller numbers. This to me is a dangerous and deeply troubling idea,and I believe a genuine appreciation for the intrinsic value of all humanbeings, not just those who are useful at the moment, will be needed toforestall it. Moreover, a good argument from future utility can also bemade: we cannot accurately predict which humans will be the great inventorsand major contributors of the future, the very people we need to addressanthropogenic climate change and many other challenges. If we value allpeople, and build a social environment in which everyone can flourish,many innovators of the future will emerge, even from unexpected quarters. <P>Threats notwithstanding, I don't think AI or machine learningcan go back into Pandora's box, and as a computer scientist who has beenproviding computing support for machine learning since long before itbecame popular, I would not want it to. AI is a powerful tool, and likeall powerful tools, it can be used for many good things. Let us build aworld together in which it is used for good, not harm.</description> </item> <item> <title>Why we thought for a while Pluto was a planet, but it never was.</title> <pubDate>Mon, 26 Aug 2019 06:51:00 -0400</pubDate> <link>http://www.cs.toronto.edu/%2525252525252525252525252525252525257Ejdd/blog.cgi/2019/08/26#pluto</link> <category>/misc</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%2525252525252525252525252525252525257Ejdd/blog.cgi/misc/pluto</guid> <description><IMG src="/~jdd/blog/misc/pluto-1315109_640.jpg" alt="Pluto" style="float:left;padding:10px"> <P> More than a decade after Pluto's demotion from the rank of planet, somestill do not accept it. I can sympathize. Like many of us, I grew upmemorizing in school the nine planets of the Solar system, the last ofwhich was Pluto: icy, distant and mysterious. I remember as a child poringover a diagram of the solar system, marvelling at the concentric elipses ofthe planetary orbits, and wondering why Pluto's orbit was so odd. For oddit was: all the other planets orbited the sun in more or less concentricelipses, but Pluto was eccentric: its orbit was at an unusual angle, andit even briefly came closer to the sun than Neptune. None of the otherplants had orbits like this: why Pluto? But I didn't question that it was aplanet. It had been recognized as a planet since Clyde Tombaugh discoveredit before my parents were born. For me, Pluto was weird, but it was still"planet", the astronomical equivalent of a sort of odd uncle who behavedstrangely and kept to himself, but still family. <P> But the idea of Pluto as a planet started to become problematic inthe early 1990s. In 1992, Jewitt and Luu discovered another object beyondNeptune: Albion, much smaller than Pluto, and also with an odd orbit. Becauseit was a small object, it was pretty clearly not a planet, so Pluto'sstatus was not yet in question, but it was only the first of many. By 2000,more than seventy such objects had been discovered. Most of these were verysmall, but some were not so small. And the discoveries continued. In 2003,with the discovery of the Eris, a trans-Neptunian body moremassive than Pluto itself, the problem became acute. No longer was Plutothe odd uncle of the planets: now there were on the order of 100 odduncles and aunts, and at least one of them, Eris, aptly named after the greekgoddess of discord, had a better claim to planethood than Pluto itself.Something had to be done. This bunch of odd objects, odd in the same way asPluto, were either all planets, or they were none of them planets. Therewas no reasonable distinction that could be made that would keep Plutoa planet but deny planethood to Eris and many of her siblings. To do sowould be arbitrary: we would be saying that Pluto was a planet simply becausewe discovered it first and it took us a long time to discover the others.What to do? <P> Happily, there was a precedent: this sort of thing had come upbefore. In 1801, Giuseppe Piazza discovered Ceres,a body orbiting between Mars and Jupiter. This was a big deal. Only twentyyears before, a new planet had been discovered for the first time inrecorded history: Uranus, found by accident by William Herschel in 1781. Now,twenty years later, Piazza had found a second. And this one was not out beyondSaturn, it was nearer than Jupiter. But Piazza's share of the limelightwas soon to lessen. his planet had a rival: a year later, Heinrich WilhelmOlbers discovered Pallas, another body between Jupiter and Mars. Twoyears later, in 1804, Karl Harding discovered another: Juno. Not to beoutdone, Olbers in 1807 discovered yet another, Vesta. By the middle ofthe 19th century, fifteen bodies orbiting between Mars and Jupiterwere known, and while none of them were anywhere as large as Ceres, oneof them, Vesta, had nearly a third of Ceres' mass. Were there really manysmall planets between Mars and Jupiter, or were these something else?When in 1846 the planet Neptune was discovered beyond Uranus, it becameclear that some decision about these bodies between Mars and Jupiter neededto be made. A consensus emerged: Ceres and other such objects were notplanets. They were called "asteroids", a name coined in 1802 by WilliamHerschel. It was a good call: there are now well over 100,000 knownasteroids, far too many for schoolchildren to memorize. <P> With Pluto, a similar situation was now occurring. While we weren'tyet at 100,000 Pluto-like bodies, we knew about quite a few more thanfifteen. And Pluto, unlike Ceres, wasn't even the most massive: Eris was,and quite possibly, bigger ones would be found. There was no denying thefacts. Pluto, like Ceres, could not be a planet. It must be something else. <P> Of course this was quite controversial. People had been calling Plutoa planet for the better part of a century. Generations of schoolchildrenhad memorized it as part of the list of planets. But the choice was clear:either the schoolchildren would have to start memorizing longer lists,much much longer ones, or Pluto would have to be demoted. Well, notdemoted, exactly, but newly recognized for what it really was allalong: something different. In the sumer of 2006, the <AHREF="http://www.iau.org">International Astronomical Union(IAU)</A> declared that Pluto isn't a planet, it is a <AHREF="https://www.iau.org/static/resolutions/Resolution_GA26-5-6.pdf"><EM>dwarfplanet</EM></A>. While this designation is a little confusing (if adwarf planet isn't a planet, why is it called a dwarf <EM>planet</EM>?),one thing was now clear: Pluto is not the same sort of thing as Mercury,Venus, Earth, Mars, Jupiter, Saturn, Uranus and Neptune; it, and Eris,and probably a couple of other larger trans-Neptunian bodies discoveredsince the 1990s, are something different. But guess what: Ceres, too,fits IAU's definition of dwarf planet, the only asteroid that does. Twocenturies after its discovery, Ceres, first-born of the non-planetsand largest of the asteroids, was deemed a dwarf planet, and Piazza, itsdiscoverer, though not the second person in recorded history to discovera new planet, was recognized as the very first to discover a dwarf one.</description> </item> <item> <title>Ross Anderson's <EM>Security Engineering</EM></title> <pubDate>Fri, 19 Jul 2019 16:13:00 -0400</pubDate> <link>http://www.cs.toronto.edu/%252525252525252525252525252525252525257Ejdd/blog.cgi/2019/07/19#anderson</link> <category>/it</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%252525252525252525252525252525252525257Ejdd/blog.cgi/it/anderson</guid> <description><A HREF="https://www.cl.cam.ac.uk/~rja14/book.html"><IMG src="https://www.cl.cam.ac.uk/~rja14/Papers/book2coversmall.jpg" alt="Security Engineering - Second Edition" style="float:left;padding:10px"></A> <P>Until recently, I had not read Ross Anderson's <EM>SecurityEngineering</EM>, despite hearing good things about it. I'm not sure why:I think I was put off a bit by the title. I had a vague and confusedimpression that a book about "Security Engineering" would be yet anotherhow-to book about making computers secure. I should have known better. Inthis case, I was wrong, very much so, and much to my detriment. I shouldhave read this book long ago. <P> Why had I not read it? I have no excuse. The book hasbeen out for a while: it is in its second edition, which cameout in 2008 (Anderson is writing a third edition, expected nextyear). So I certainly had the opportunity. Moreover, since 2012,the book has been free for the reading (and downloading) from <AHREF="https://www.cl.cam.ac.uk/~rja14/book.html">his website</A>. So Icertainly had the means. I just didn't, until a few weeks ago, when Istumbled across it again. I read a little from the website, then a littlemore. Before long, I was well and thoroughly hooked. <P><EM>Security Engineering</EM> is a classic, comprehensive book aboutinformation security: eminently readable, clear and thorough, it coversinformation security in pretty much every aspect one might encounter it,from the usual (cryptography, access controls, protocols, biometrics) to thenot quite so day-to-day (nuclear weapons launch protocols, counterfeiting,even spying by analyzing the RF emissions from computers). Each chapteris a clear elucidation of a particular aspect of information security,focusing on the essential issues. Each chapter provides enough detail tounderstand the essential elements, yet not too much detail as to overwhelmthe reader. His writing is a classic illustration of the difference betweenan expert and a master. An expert knows a great deal about a topic andprovides an abundance of information. A master knows the key elements,those things that are most important, on which everything else hangs, andfocuses exactly on these. This book is mastery, in clear, understandableand engaging language. It has become my favourite book in informationsecurity already, and I haven't yet finished it. <P>I look forward to the third edition sometime next year. I can't wait.</description> </item> <item> <title>Externality and Information Security</title> <pubDate>Mon, 04 Mar 2019 12:04:00 -0500</pubDate> <link>http://www.cs.toronto.edu/%25252525252525252525252525252525252525257Ejdd/blog.cgi/2019/03/04#externality</link> <category>/it</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%25252525252525252525252525252525252525257Ejdd/blog.cgi/it/externality</guid> <description>It was a hot midsummer weekend, and I was traveling back to Torontowith friends. We were on the expressway (the name here in Ontario forthe sort of road that Americans call freeways and Brits call motorways).Traffic was very slow: a classic traffic jam. After about thirty minutes,we reached the cause of the problem. It was not a collision. Nor was ithighway construction. Instead, by the side of the roadway, a minivan wasparked, back gate open, and a family was having a picnic on the nearbygrass. I don't know if they realized they were causing a traffic jam, butthey were. People had slowed to look, which caused traffic behind to slowtoo, and because of the traffic volume, this led to a traffic jam over aconsiderable distance. <P> I don't know why the family having the picnic had chosen that spot forit, and I don't know whether they realized the problem they were causing.But their picnic went on, unaffected by the traffic problems they werecausing. In other words, the traffic jam was <em>not their problem</em>.It was an <em>externality</em>, something causing a negative effect notfelt by those who cause it. <P> Externalities happen in life all the time. Large organizations(companies, countries, institutions) suffer significantly when theirdecision-makers make decisions that are good for themselves but not goodfor the organization. Rules to make this less likely are put in place:rules against bribery, rules concerning conflict of interest, rules imposingdue process. But rules only work to a certain extent: there are plenty ofsituations where the rules are followed yet still externalities happen.Moreover, rules come with costs, sometimes significant ones. Rules maybe necessary, but they are not sufficient, and they need to be accompaniedby buy-in. <P>Let's consider traffic again. Driving is governed by all sorts of rules.Some of these rules work well: at traffic lights, go when the light isgreen, stop when it is red. Rarely broken, this rule makes traffic workin dense situations where otherwise there would be chaos. Most of thetime, this rule is followed even in the absence of external enforcement.When enforcement does occur, it is well regarded: hardly anyone will arguethat a person running a red light is a safety hazard and should be ticketed.In practice, you can stand for hours beside a busy traffic signal in atypical Ontario city, and despite the absence of police presence, notfind a single driver running a red light. <P>Sadly, other driving rules don't work quite so well, such as speed limitson expressways here in Ontario. These limits are often broken, with somefollowing them and others not. Often, on an uncongested expressway,unless enforcement is likely (i.e. police is present) there will be somepeople driving over the speed limit. Enforcement is viewed cynically:speeding tickets are often viewed more as revenue generation than as asafety measure. Obeying speed limits is often viewed by drivers asan externality: <em>not my problem</em>, unless there is a police officeraround to make it one. In practice, at any place on any uncongestedOntario expressway, you will be hard-pressed to find a five-minute periodin which no passing driver has exceeded the speed limit. <P>I have been thinking a lot about information security lately. Ininformation security, we have a situation similar in many respects todriving. Just as driving is a matter of traveling safely, informationsecurity is a matter of computing safely. When we compute, we maybe processing information that is sensitive, confidential, private. Harmcan occur when it is exposed. Steps need to be taken to ensure that itis not: persons handling information will have to handle it securely. Butdo we want this process to look like speed limits? Or traffic lights? Ithink the answer is clear: if we want information to actually be secure,we want good security practice to be followed like the rules for trafficlights are followed: broadly and consistently, without the need for theconstant threat of enforcement. <P>In recent years, an information security profession hasarisen. The increasing demands of the profession have made it increasinglyrare that an information security professional has spent much time actuallyrunning a substantial IT operation. Certifications abound, and amultiplicity of complex and large security standards have been created,each requiring professionals to interpret. A great deal of money isbeing spent on information security. Much of this is good and necessary:information security needs attention, codification, dissemination, andchampionship. But the professionalization of information security comeswith big risks, too: the risk that information security will become theresponsibility only of specialists, the risk that these specialists willcome up with all-encompassing codexes of security standards to impose,the risk that these standards will be treated as externalities by ITpractitioners, the risk that the information security profession willrespond with enforcement, and hence the risk we will find ourselves inthe expressway speed limit situation with respect to information security. <P>The fact is, information security is an aspect of good IT practice:if an implementation is not secure, it is <em>broken</em>, just as muchas if it were not reliable. Security is the responsibility of all ITpractitioners: it needs to be internalized, not externalized. <P>For this to happen, it is important that information security rulesbe simple and understandable, to ensure buy-in. Just as traffic lightrules address the obvious risk of traffic accidents, so should securityrules address clear risks in a visibly appropriate way. In most cases,it's not so important that rules be part of a comprehensive codex thataddresses all possible areas of risk: the more complex the rule and themore extensive the system of rules, the more likely it will all be treatedas an externality. What we really want are not rules for their own sake,but genuinely secure IT. <P> If we want secure IT, we need to recognize that there is anotherpotential externality at work. Genuine information security and the good ofthe information security profession may not always align. Just as expresswayspeed limits employ more police than traffic lights, an enforcement approachwill employ more information security professionals than an internalizedone. But the internalized approach is what gives us secure computing. Thisis not something that can be left to the information security professionalone. To get there, we will need collaborative effort from all of us,particularly those with long experience running substantial IT operations.We will all need to make a true commitment to a practical approach, onethat seeks to make computing genuinely more secure in the real world.</description> </item> <item> <title>Intentionality</title> <pubDate>Tue, 26 Feb 2019 06:27:00 -0500</pubDate> <link>http://www.cs.toronto.edu/%2525252525252525252525252525252525252525257Ejdd/blog.cgi/2019/02/26#intentionality</link> <category>/misc</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%2525252525252525252525252525252525252525257Ejdd/blog.cgi/misc/intentionality</guid> <description><p> I spent all of 2018 intending to blog, and not doing it. Sadly, thisis an all too human situation. We intend to do things, when we can, whentime permits, but we can't; time doesn't permit. Or at least this isone of those stories we tell ourselves. The truth is a little simpler:throughout 2018, my intention to blog was not strong enough for me tore-prioritize things in my day so that I would do it. <p> I had plenty to say. I continue to have plenty to say. I had plentyof important things to do, and that also continues to be true. Despite myother responsibilities, I am making time now, and I will continue tomake time, every so often, to say things in this blog. I am being<em>intentional</em> about it. <p> To be intentional about something means to be deliberately purposeful:to make one's actions a directly chosen consequence of one's thoughtfuldecisions. For most people, myself included, life is full of input,distractions, demands, requests. It is easy to fill time without mucheffort. But if I am not intentional, it will be filled with reaction,not action: things that circumstances and prior commitments have chosenfor me, not things I have chosen for myself. <p> Reaction is fine, even good and necessary. Many people, myself included,build up throughout their lives various important responsibilities:responsibilities to family, work, friends, communities. Responsibilitiescarry with them a commitment to react to the needs of others. This iswell and good. But it is not enough, at least not for me. I realize thatto be authentic, I have to consider carefully what is important to me,decide what to do about it, and then act on it. This is intentionality. I'vedecided to be intentional about blogging. Look for moreblog entries in the coming weeks.</description> </item> <item> <title>A Way to Visualize Relative Masses of Things in the Solar System </title> <pubDate>Tue, 12 Dec 2017 13:07:00 -0500</pubDate> <link>http://www.cs.toronto.edu/%252525252525252525252525252525252525252525257Ejdd/blog.cgi/2017/12/12#astromass</link> <category>/misc</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%252525252525252525252525252525252525252525257Ejdd/blog.cgi/misc/astromass</guid> <description>Every so often we hear things in the news about the solar system: a missionto a planet or asteroid, talk of manned missions to mars, arguments aboutwhether Pluto is a planet or not. We tend to have pretty sketchy ideas ofwhat most bodies in the solar system are like compared to Earth. The factis that they're more wildly different in size and mass than we might think. <P>Let's look at mass. Imagine you decide to row across San Franciscobay in a 12-foot aluminum rowboat. You pack a couple of suitcases, your15 inch Macbook Pro (can't go without connectivity) and your ipad mini,you get in your rowboat and start rowing. As you row, you get hungry, soyou pull out a Snickers bar. Now imagine that the USS Nimitz, a massivenuclear-powered aircraft carrier, passes by. There you are, in a rowboatwith your two suitcases, your Macbook Pro, your iPad, and your Snickers bar,alongside a huge supercarrier. <P> Well, the mass of the sun compared to the earth is like that aircraftcarrier compared to you and your boat. The mass of Mars is like yourtwo suitcases. The mass of the moon is like your 15 inch Macbook Pro,and the mass of Pluto is like your iPad mini. As for the Snickers bar,it's like Ceres, the largest of the asteroids. <P>Now let's suppose the massive wake of the aircraft carrier tips overyour rowboat and leaves you in the water. Along comes a rich tech founderin his 70 foot yacht, and fishes you out. That yacht is like Jupiter,the largest planet. <P>So forget any mental images you might have of planets being somethinglike the Sun, only a bit smaller and cooler. The sizes of things in thesolar system are really quite different, and there is nothing, absolutelynothing, in the solar system that is anything quite like the Sun.</description> </item> <item> <title>Bitcoin, Cryptocurrency and Blockchain</title> <pubDate>Mon, 11 Dec 2017 14:02:00 -0500</pubDate> <link>http://www.cs.toronto.edu/%25252525252525252525252525252525252525252525257Ejdd/blog.cgi/2017/12/11#bitcoin</link> <category>/it</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%25252525252525252525252525252525252525252525257Ejdd/blog.cgi/it/bitcoin</guid> <description><P>As the price of Bitcoin goes up and up, talk increases about Bitcoinand other cryptocurrencies, like Litecoin, Monero, ZCash, Ethereum,and many others. Plenty is being said, and it can be a bit confusing. <P>But there is no need to be confused. Bitcoin and other cryptocurrenciesare basically simple. They are not coins. They are simply lists.Each cryptocurrency has a master list. The list typically containsinformation about <em>who</em> and <em>what</em> (i.e. amounts). The listis designed in a clever way, using computer software, so that peopleall over the world can have identical copies of the list and keep it upto date, without someone having to be the holder of the "master copy".But it is still just a list. <P> The sort of list used for cryptocurrencies is called a "blockchain",and it has some special properties. One particularly clever property isthat you can't normally just add anything you want to the list, there is ascheme to control that. Instead, you need to arrange with someone alreadyon the list to give up (some of) their place on the list to you. <P>So when someone says they bought some Bitcoin and they're going to makea lot of money, what they mean (whether they realize it or not) is thatthey paid somebody some money to put them on a list, and they hope thatsomeone later will pay them even more money to get off it. <P>As for me, I haven't "bought" any. As I write this, cryptocurrencyprices are rising fast. But I think what is happening is a kind of run-awaypositive feedback loop: people are buying in because it is going up, and itis going up because people are buying in. Eventually it will run out ofpeople to buy in, and it will stop going up. Then some people will sell,causing the feedback loop to go the other way: people will sell becauseit is going down, and it will go down because people are selling. <P>That being said, one thing in particular about cryptocurrency ismaking me grumpy about it, even though I don't "own" any. Recall I wrotethat you can't <em>normally</em> make yourself a new entry on a blockchainlist, but there is a way. You can do an enormous lot of computations ona computer in an attempt to find new special numbers that can be usedto create new entries on the list. This process is misnamed "mining",but it's more a sort of computerized brute-force mathematical searching.Those computations take a long time and use a lot of electricity. Moreover,even the ordinary transactions generated by people "buying" and "selling"a cryptocurrency is a computational burden, since thereare so many copies of the list around the world. Each listis very big: Bitcoin's is more than 100GB, and every copyneed to be updated. This uses electricity too. In fact, <AHREF="https://digiconomist.net/bitcoin-energy-consumption">digiconomist.netestimates that Bitcoin computations alone presently use up enoughelectricity to power more than three million US households.</A>Furthermore, the "mining" computers use GPUs that arereally good for graphics and machine learning, but because cryptocurrency"miners" are buying them all up, those GPUs are getting harder tofind for a good price. Personally, I am not happy with thechallenges I am having in finding enough GPU resources for our computerscientists, who are hungry for GPUs for machine learning. While highdemand for GPUs is maybe good for GPU manufacturers (for example, <AHREF="http://fortune.com/2017/09/03/bitcoin-graphics-card-makers-stocks/">accordingto fortune.com, Nvidia made U$150M in one quarter in 2017 selling GPUsto cryptocurrency "miners"</A>), surely all those GPUs, and all thatelectricity, can be used for something more useful than cryptocurrency.</description> </item> <item> <title>A closer look at topuniversities.com's 2017 rankings for Computer Science.</title> <pubDate>Thu, 09 Mar 2017 12:58:00 -0500</pubDate> <link>http://www.cs.toronto.edu/%2525252525252525252525252525252525252525252525257Ejdd/blog.cgi/2017/03/09#rankings2017</link> <category>/misc</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%2525252525252525252525252525252525252525252525257Ejdd/blog.cgi/misc/rankings2017</guid> <description><p><A HREF="http://topuniversities.com">The QS World University Rankings</A>for 2017 are out, including the subject rankings. For the subject "ComputerScience & Information Systems", the University of Toronto does very well,placing tenth. <P>A closer look at the top ten shows some expected leaders (MIT, Stanford,CMU, UC Berkeley) but some less expected ones, such as Oxford and Cambridge.These are superb Universities with good Computer Science programs, butare their CS programs really among the ten best in the world? <P>A closer look at how the score is computed sheds some light on thisquestion. The Overall Score is a combination of Academic Reputation,Citations per Paper, Employer Reputation, and H-index Citations.Academic Reputation and Employer Reputation are, in essence, the opinions ofprofessors and employers respectively. While (hopefully) they are reasonablywell founded opinions, this is a subjective, not an objective, metric.On the other hand, Citations per Paper and H-index Citations are objective.So I looked at Citations per Paper and H-index Citations for the top fortyschools on the 2017 QS Computer Science & Information Systems ranking. <P>By Citations per Paper, top five of those forty are:<OL><LI>Princeton<LI>Stanford<LI>UT Austin<LI>Washington<LI>UC Berkeley</OL> <P>No MIT? This seems off. So lets look at the top five by H-Index Citations: <OL><LI>Stanford<LI>MIT<LI>UC Berkeley<LI>UI Urbana-Champaign<LI>UT Austin</OL> <P>That looks more reasonable. So let's look at the top twenty by H-Index Citations:<OL><LI>Stanford<LI>MIT<LI>UC Berkeley<LI>UI Urbana-Champaign<LI>UT Austin<LI>Georgia IT<LI>CMU<LI>Tsinghua<LI>Nanyang<LI>ETH Zurich<LI>Washington<LI>Princeton<LI>UBC<LI>Toronto<LI>Waterloo<LI>NU Singapore<LI>UC London<LI>Cornell<LI>UCLA<LI>CU Hong Kong</OL> <P>That's a list that makes more sense to me. While it puts my department14th instead of 10th, I think I have more confidence in the objectivityof this ordering than I do in the QS Overall Score ordering.</description> </item> <item> <title>Program Source Code Should be Readable by Human Beings By Definition</title> <pubDate>Thu, 02 Feb 2017 13:35:00 -0500</pubDate> <link>http://www.cs.toronto.edu/%252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/2017/02/02#python3</link> <category>/it</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/it/python3</guid> <description>Version 3 of the <A HREF="https://www.python.org">Python programminglanguage</A> made a seemingly innocuous change to the Python programminglanguage: <A HREF="https://www.python.org/dev/peps/pep-0008/#id18">nolonger could tabs and spaces be mixed for indentation: either tabs must beused exclusively, or spaces.</A> Hence the following is not a valid Python3 program:<pre>def hello(): print("Hello") print("World")hello()</pre>If I run it, here's what I get:<pre>% python3 testme.py File "testme.py", line 3 print("World") ^TabError: inconsistent use of tabs and spaces in indentation</pre>However, the following is a valid Python 3 program:<pre>def hello(): print("Hello") print("World")hello()</pre><pre>% python3 testme.pyHelloWorld</pre>and so is the following:<pre>def hello(): print("Hello") print("World")hello()</pre><pre>% python3 testme.pyHelloWorld</pre> Confused yet? <p>As you can, or perhaps more to the point, <em>can't</em> see, the problemhere is that the first program uses a tab to indent the first<em>print</em> statement, and spaces to indent the second <em>print</em>statement. The second program uses spaces to indent both, and the thirdprogram uses tabs to indent both. But because tabs and spaces are bothvisually represented as whitespace, it is difficult or impossible tovisually distinguish between a correct and an incorrect python3 programthrough inspecting the source code. This breaks the basic definition ofsource code: <em>human-readable computer instructions</em>. <P>No doubt the Python 3 designers have good intentions: to helppython programmers be consistent about indentation. But to me, itseems unreasonable to have a programming language where syntacticallyor semantically important distinctions are not clearly visible in thesource code.</description> </item> <item> <title>Slow Windows Update on Windows 7 again? Install two Windows Update patches first.</title> <pubDate>Wed, 23 Nov 2016 09:48:00 -0500</pubDate> <link>http://www.cs.toronto.edu/%25252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/2016/11/23#wupdate2</link> <category>/it</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%25252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/it/wupdate2</guid> <description>Back in May, <AHREF="http://www.cs.toronto.edu/~jdd/blog.cgi/2016/05/09#wupdate">I wrote</A> about Windows Update for Windows 7 taking many hours or evendays; the fix then was to install two patches manually first. <P> The problem has returned. Even if you install the two patches I mentionedin May, you may experience very slow updates on Windows 7. <P>Happily, again there's a workaround: grab two patches,different than before, and manually install them. Get <AHREF="https://support.microsoft.com/en-ca/kb/3172605">KB3172605</A> and its prerequisite <AHREF="https://support.microsoft.com/en-ca/kb/3020369">KB3020369</A> from the<A HREF="https://www.microsoft.com/download">Microsoft Download Center</A>,and install them manually in numeric order, before running Windows update.If making a fresh Windows 7 installation, simply install Windows 7 SP1,followed by KB3020369, then KB3172605, and only then run windows update.These two patches seem to address the slowness issues: after they wereinstalled on some of our systems here, Windows Update ran in a reasonableamount of time.</description> </item> <item> <title>Remembering Kelly Gotlieb</title> <pubDate>Wed, 26 Oct 2016 10:41:00 -0400</pubDate> <link>http://www.cs.toronto.edu/%2525252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/2016/10/26#gotlieb</link> <category>/misc</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%2525252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/misc/gotlieb</guid> <description><P>On October 16th, 2016, <AHREF="http://web.cs.toronto.edu/news/current/In_Memoriam__The__father_of_computing_in_Canada__Calvin_C__Gotlieb.htm">KellyGotlieb</A>, founder of the <A HREF="http://www.cs.toronto.edu">Departmentof Computer Science</A> at the <A HREF="http://www.utoronto.ca">Universityof Toronto</A>, passed away in his 96th year. I had the privilegeof knowing him. Kelly was a terrific person: brilliant, kind, andhumble. He was always willing to make time for people. He was a greatthinker: his insights, particularly in the area of computing and society,were highly influential. I never fully realized how influentialhe was until we, here at the department of Computer Science,created a <A HREF="http://socialissues.cs.toronto.edu">blog,http://socialissues.cs.toronto.edu</A>,in honour of the 40th anniversary of <em><A HREF="https://books.google.ca/books?id=xLTiBQAAQBAJ">SocialIssues in Computing</A></em>, the seminal textbook he and <AHREF="http://www.cs.toronto.edu/~bor">Allan Borodin</A> wrote in 1973in the area of computers and society. I served as editor of the blog,and solicited contributions from the top thinkers in the field. So manyof them responded, explaining to me how influential his ideas had been tothem, and the blog was filled with insightful articles building in variousways upon the foundation that he and Allan had laid so many years before. Iinterviewed Kelly for the blog, and he was terrific: even in his nineties,he was full of insights. His mind active and enthusiastic, he was makingcogent observations on the latest technologies, ranging from self-drivingcars to automated medical diagnosis and treatment. <P>To me, Kelly epitomized the truth about effective teaching that is all toooften missed: teaching is not just about information, teaching is aboutinspiration. Kelly was a truly inspiring teacher and thinker. He wascompletely authentic in everything he did, he was full of enthusiasm,and that enthusiasm was infectious. Conversations with Kelly so oftenleft me energized and inspired, thinking along new directions of thoughtthat something he said had triggered, or leaping past obstacles thathad previously seems insurmountable. That is true teaching. Informationwithout inspiration is simply fodder for forgetfulness, but teaching thatinspires leads to new insights, integration of ideas, genuine understanding,and a better, clearer and sharper window on the world. Kelly inspired somany people for so many years. We are truly blessed that he was among us.He will be remembered.</description> </item> <item> <title>The Price of Google</title> <pubDate>Sun, 16 Oct 2016 18:02:00 -0400</pubDate> <link>http://www.cs.toronto.edu/%252525252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/2016/10/16#google</link> <category>/it</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%252525252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/it/google</guid> <description>I am a Canadian still living in the city in which I wasborn. I love living in Canada, but life in Canada has its price.<A HREF="http://thecanadianencyclopedia.ca/en/article/al-purdy/">AlPurdy</A>, the late 20th century Canadian poet, once wrote about Canadaas a country where everyone knows, but nobody talks about, the fact thatyou can die from simply being outside. It is true, of course: almosteverywhere in Canada, the winter is cold enough that a sufficient numberof hours outside without protection can lead to death by exposure. Butthis basic fact is designed into pretty much everything in Canadian life,it is simply accepted as a given by well over thirty million Canadians,and we cope: we wear the right winter clothes, we heat and insulate ourbuildings in winter, we equip our cars with the right tires, and life goeson. Despite the Canadian winter, Canada is a great place to live. <P>Google offers a lot of very good free web services: it is "a greatplace to live" on the Internet, and their services are used by hundredsof milliions of people all over the world. While Google seems about asfar removed from a Canadian winter as you can imagine, there's somethingin their Terms of Service that people seem to rarely talk about, somethingthat might have a bit of a chilling effect on one's initial ardor. <P>Google, to its credit, has a very clear and easy-to-read <AHREF="https://www.google.com/intl/en/policies/terms/">Terms ofService document</A>. Here's an excerpt from the version of April 14, 2014, which is the most current version at the time I write this. <blockquote>When you upload, submit, store, send or receive content to or through ourServices, you give Google (and those we work with) a worldwide license touse, host, store, reproduce, modify, create derivative works (such as thoseresulting from translations, adaptations or other changes we make so thatyour content works better with our Services), communicate, publish, publiclyperform, publicly display and distribute such content. The rights you grantin this license are for the limited purpose of operating, promoting, andimproving our Services, and to develop new ones. This license continueseven if you stop using our Services (for example, for a business listingyou have added to Google Maps).</blockquote> Let me pull out for closer examination the most important bits.For readability, I've omitted elipses. <blockquote>When you submit content to our Services, you give Google (and those we workwith) a worldwide license to use such content for the purpose of our Services. This continues even if you stop using our Services.</blockquote> <P>As you can see, this is pretty broad. You are granting Google and theirpartners the right to use your content for Google's Services (present andfuture) anywhere in the world, forever. While it does say that it must beused for the purpose of their Services, it doesn't limit itself toexisting Services and it doesn't constrain what a "Service" might be. Sincedeveloping and offering Services, broadly understood, pretty much coversthe gamut of what Google does as a company, the answer is <em>Yes</em>:by submitting content to their services, you are granting Google and theirpartners the right to use your content anywhere in the world, forever,for a broadly unconstrained set of purposes. <P>So does this mean nobody should use Google? Does the Canadian wintermean that nobody should live in Canada? After all, as Al Purdy writes,in Canada you can die from simply being outside. <P>Well, no, of course not. While Google has the right to do broadlyunconstrained things with our content that we submit to them, their self-interest is typically aligned with our's: they want us to entrust ourcontent to them, because they use it to earn money to operate. Therefore,to persuade us to keep submitting content to them, they will work hardto protect and secure the content they already have, in ways they thinkwe consider important. For this reason, I think it's not unreasonable totrust Google with some of my content: I believe they are likely to protectit in sensible ways. Other content I choose not to submit to Google.Just as I am prepared for a Canadian winter, knowing it is the priceI pay to live in Canada, I continue to use some Google services, knowing thatthey will keep and use my content. Many Google services are very good andwell worth using, much of my content is not very sensitive, and I trustGoogle enough to share content with them. <p>I do wonder, however, how many Google users really understand the rightsthey are granting to Google. Canada has been around for centuries: theCanadian winter is no secret. But the implications of Google's broadright to use our content are not quite so obvious. It's not really so clearhow Google is using the content or might use it in the future, and even ifwe trust Google, can we trust all those who might put pressure on Google?Quite frankly, we really don't know yet how Google's massive repository ofour collective content can be used. We can envision wonderful outcomes:historians a century or two hence coming to insightful conclusions aboutearly twenty-first century society, for example, but we can also envisionoutcomes not quite so sanguine: for example, a twenty-first centuryversion of Orwell's <em>1984</em>, a dystopian world of "thought-crimes"and "doublespeak" where content is is scanned for dissent from a prevailingideology. A certain degree of caution may be warranted: in the case ofGoogle, unlike Canada, we may not have yet seen how severe winter can be.A certain degree of caution is warranted. Yes, use Google, but use itknowing what you are doing. <P>One last thing to be said: I focus on Google here, butthe same issues hold for Facebook, Twitter, Yahoo and other purveyorsof free services over the Internet. Read their Terms of Service tolearn what rights you are granting by your use of their services,and decide on the basis of that knowledge how to use their services,and even whether you use their services at all. After all, <AHREF="https://en.wikipedia.org/wiki/Snowbird_(person)">even Canadianssometimes choose to spend winter in Florida, Mexico, or Arizona</A>.</description> </item> <item> <title>The Sun-Managers Mailing list: a Knowledge Sharing Success Story</title> <pubDate>Mon, 16 May 2016 20:29:00 -0400</pubDate> <link>http://www.cs.toronto.edu/%25252525252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/2016/05/16#sunmanagers</link> <category>/it</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%25252525252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/it/sunmanagers</guid> <description><A HREF="http://sunmanagers.cs.toronto.edu">Sun-Managers</A> was an emailmailing list for system administrators of computers made by <AHREF="http://en.wikipedia.org/wiki/Sun_Microsystems">Sun Microsystems,Inc</A>. The list operated from mid-1989 to the fall of 2014, and I wasprivileged to be part of it for almost all of its history. Sun-Managers was founded in May of 1989 by <AHREF="https://www.linkedin.com/in/william-lefebvre-936b76">William (Bill)LeFebvre</A>, at <A HREF="http://www.northwestern.edu">NorthwesternUniversity</A>. At the time, Bill ran <AHREF="http://ftp.lanet.lv/ftp/sun-info/sunflash/1990/Aug/20.10.sunspots">Sun-Spots</A>,a digest-format mailing list for system administratorsof Sun systems, but the digest format made it difficultfor people to ask questions and get a timely response. He <AHREF="https://groups.google.com/forum/#!msg/comp.sys.sun/H9z8YwunvB4/xJFp8qOLTFUJ">createdSun-Managers</A>, an unmoderated mailing list intended forshort-turnaround time questions. This was an immediate success:so much so that by the fall of 1989, the sheer number of messages onthe list were swamping mailboxes. In Nov 1989, Bill instituted a <AHREF="http://sunmanagers.cs.toronto.edu/policy.txt">simple policy</A>:if someone asks a question on the list, other list members were expectedto reply by email directly to the person asking the question, not to thelist. The person asking the question, in turn, was expected to summarizethe answers received, and send the summary to the list. <P> I joined the list about this time: I had started a new job atthe <A HREF="http://www.utoronto.ca">University of Toronto</A>'s <AHREF="http://www.cs.toronto.edu">Computer Science department</A>, a role thatincluded the administration of a number of Sun workstations and servers.I was looking for resources to help me with my Sun system administrationtasks, and this list was an excellent one. Because of this summary policy, thelist volume was manageable enough that I could keep up, yet the turnaroundtime on questions was short. I mostly "lurked" at first, reading but notreplying. I felt too inexpert to answer many questions, and too shy to ask.However, I learned a great deal from what I read. Moreover, the summarieswere archived, and <A HREF="http://sunmanagers.cs.toronto.edu/year.html">this archive</A> became a resource in itself, a knowledge-base of practicalinformation about administering Sun systems. <p> The list grew very rapidly: 343 summaries in 1990, and over 1000 in1991. In August of that year, it was noted that certain questions were beingasked often, and rather than waste effort answering the same question severaltimes, a <A HREF="http://sunmanagers.cs.toronto.edu/faq.html">"FrequentlyAsked Questions" (FAQ) file</A> was instituted. The <AHREF="http://sunmanagers.cs.toronto.edu/1991/0484.html">first version</A> was created by a list member from <A HREF="http://www.bu.edu">BostonUniversity</A>, and quickly grow to dozens of answers. <p> By November of 1992, the list had grown to thousands of members,and the workload of managing the list, editing the FAQ and coachinglist members on how to follow the <A HREF="http://sunmanagers.cs.toronto.edu/policy.txt">list policy</A>had become significant. Many list members were not individuals,but "mail exploders": email addresses that themselves were mailinglists going to multiple individuals at a given site. This madehandling list membership issues more complex. Bill LeFebvre decidedto hand the list over to others. Two list members stepped up: <AHREF="http://www.linkedin.com/in/gene-rackow-87b7b25">Gene Rackow</A> from<A HREF="http://www.anl.gov">Argonne National Laboratory</A> to run thelist software, and me, to handle the FAQ and policy work. By this time,I had benefitted from the list for a while, and I felt it was time to"give back". At the time, I wasn't in a position to actuallyrun the list: I'd just taken on a new role as system manager ofthe <A HREF="http://www.utoronto.ca">University of Toronto </A><AHREF="http://www.cs.toronto.edu">Computer Science Department</A>'s <AHREF="http://www.cdf.toronto.edu">teaching laboratories</A>, and had myhands full, but I could certainly help with content. I was really glad towork together with Gene, a seasoned system administrator, on this rapidlygrowing list, which we moved to a system at <AHREF="http://www.anl.gov">Argonne National Labs</A>, where Gene worked. <p> The list continued to grow through the 1990s. During this time,Sun Microsystems was quietly supportive, helping Gene with hardware (a<A HREF="http://en.wikipedia.org/wiki/SPARCstation_1">Sparcstation 1</A>)as the list grew. By 1996, over two thousand summaries a year were beingproduced, peaking at 2243 in 2002. In May of 1998, Gene Rackow handed over list management to RobMontjoy from the University of Cincinnati, who in turn handed overlist management to <A HREF="https://www.linkedin.com/in/mrbill">BillBradford</A> in November of 2000. The list was moved from ArgonneNational Labs to a system in Austin run by Bill. I continued to managethe <A HREF="http://sunmanagers.cs.toronto.edu/policy.txt">listpolicy</A> and edit list information files, such as <AHREF="http://sunmanagers.cs.toronto.edu/before-posting.txt">a"think before posting" reminder</A> and the FAQ which had grown to79 questions by December 2000. This had become a bit too large, andso <A HREF="http://sunmanagers.cs.toronto.edu/faq-2000.txt">19questions deemed less frequently asked</A> weretrimmed. A further trim was made in 2005, reducing a <AHREF="http://sunmanagers.cs.toronto.edu/faq-2005.txt">65-question FAQ</A>to <A HREF="http://sunmanagers.cs.toronto.edu/faq.txt">one under 60</A>. <p> By 2002, the list had reached over five thousand members and theworkload of running the list software and managing the listsubscriptions had become too much for one person. Dan Astoorian, mycolleage at the <A HREF="http://www.utoronto.ca">University of Toronto</A>,stepped in to help, and he was sorely needed. Moreover, the list serverhardware was feeling the strain: by mid-2001, list members were beingasked to contribute used equipment to upgrade the server. This wasresolved in April 2003, when the list was migrated to a machine at the<A HREF="http://www.utoronto.ca">University of Toronto</A> that had beendonated to the University by Sun Microsystems. <p> But times were changing. Linux was growing rapidly and Sun's business was being affected. The web provided more resources for people seekinghelp administering their systems, and fewer were relying on mailing lists.The list fell below 2000 summaries per year in 2003, under 1200 in 2004, and dropped below 1000 in 2005. By 2008, summaries per year had fallen to about 300, fewer than in any full-year period previously. Sun Microsystems ran into significant difficulties during the economic downturn that year, and was sold to<A HREF="http://www.oracle.com"> Oracle</A> the following year. As for the list, in 2009, there were just over 200 summaries, declining to less than 100 in 2011. More disturbingly, the ratio of summaries to questions was steadily declining,from over 24% in 2001 to less than 16% by 2010: for some reason, list members were becoming less diligent in summarizing responses back to the list. Summaries and list traffic in general continued to decline rapidly: there were just over 50 summaries in 2012, and less than a dozen in 2013. In 2014, there were only three by October, when a hardware failure provided a good excuse to retire the list. <p> The Sun-Managers mailing list, over its twenty-five year lifetime, provided help to many thousands of system administrators, producing over29000 summaries, <A HREF="http://sunmanagers.cs.toronto.edu/year.html">anarchive of which continues to be available</A>. Special thanks is due tothe superb people I was privileged to work together with on the list overthe years: William LeFebvre, Gene Rackow, Rob Montjoy, Bill Bradford, andDan Astoorian. Gratitude, also, is due to the thousands of list memberswho so freely shared their knowledge and expertise with others. <P> The list summary archive, and an account of the list'shistory (on which this blog entry is based) is available at <AHREF="http://sunmanagers.cs.toronto.edu">http://sunmanagers.cs.toronto.edu</A>.The list's official web page, <AHREF="http://www.sunmanagers.org">http://www.sunmanagers.org</A>, continues to bemaintained by Bill Bradford.</description> </item> <item> <title>Slow Windows Update on Windows 7? Install two Windows Update patches first.</title> <pubDate>Mon, 09 May 2016 10:54:00 -0400</pubDate> <link>http://www.cs.toronto.edu/%2525252525252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/2016/05/09#wupdate</link> <category>/it</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%2525252525252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/it/wupdate</guid> <description>Recently, I noticed Windows Update taking many hours or even days onWindows 7, especially for new installs/reinstalls. Task manager showssvchost.exe exhibiting large memory usage (suggestive of a memory leak)and/or sustained 100% CPU. <P>Happily, there's a workaround: grab a couple of patchesto Windows Update itself, and manually install them. Get <AHREF="https://support.microsoft.com/kb/3050265">KB3050265<A> and <AHREF="https://support.microsoft.com/kb/3102810">KB3102810</A> fromthe <A HREF="https://www.microsoft.com/download">Microsoft Download Center</A>,and install them manually in that order, before running Windows update.These two patches seem to address the issues: after they were installedon some of our systems here, Windows Update ran in a reasonable amount oftime (an hour or two perhaps on slow systems when many updates are needed,but not days).</description> </item> <item> <title>Apple vs FBI: it is about setting a precedent.</title> <pubDate>Fri, 04 Mar 2016 10:25:00 -0500</pubDate> <link>http://www.cs.toronto.edu/%252525252525252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/2016/03/04#applefbi</link> <category>/it</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%252525252525252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/it/applefbi</guid> <description>There seems to be lots of confusion about Apple's current dispute withthe FBI, despite <A HREF="http://www.apple.com/customer-letter">Apple'smessage to their customers of Feb 16, 2016</A>, where they tried toexplain the issue. Here's the issue in a nutshell. <p> The FBI has an Apple iPhone that was the work-phone of a now-deadterrorist. The FBI wants to read what is on that phone. But thephone is encrypted, and runs a secure version of iOS. The FBI wants Appleto make an insecure version of iOS to run on that phone, so that the FBIcan break into the phone and read the contents. Apple has, so far, refused. <p> This issue will no doubt be addressed in the US courts and legislatures.What is at stake is the precedent it sets. The essential question is this:to what extent should law enforcement be able to compel others to assistthem with an investigation? Should software developers be expected tomake insecure versions of their software, so that law enforcement can"break in"? It will be very interesting to see how this plays out.</description> </item> <item> <title>Apple's new Macbook laptop: like a tablet?</title> <pubDate>Fri, 13 Mar 2015 11:08:00 -0400</pubDate> <link>http://www.cs.toronto.edu/%25252525252525252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/2015/03/13#macbook</link> <category>/it</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%25252525252525252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/it/macbook</guid> <description><P>I rarely write about Apple's products because they have no shortage ofpress already: Apple has superb marketing, and many of their products areremarkable in one way or another, often for excellent design and engineering.Their new super-thin <A HREF="http://www.apple.com/macbook/">Macbooklaptop</A> is no exception: it's very thin and light, has a superbhigh-resolution screen, a carefully redesigned trackpad and keyboard,and is very power-efficient. New to this machine is the fact that ithas only a single USB-C port for power, data, and video (it also has aheadphone port for audio). Most laptops have many more ports than this.A USB port used for both power and data, and a headphone port, but nothingelse, is more typical of a tablet, not a laptop. Indeed, some of the press seems to havereally latched onto this "tablet" comparison. <AHREF="http://www.foxnews.com/tech/2015/03/11/apple-has-designed-macbook-thats-almost-tablet/">BrookeCrothers of Foxnews/Tech claims that the MacBook is "almost a tablet"and states that the MacBook "is an iPad with a keyboard" </A>while <AHREF="http://www.slate.com/blogs/future_tense/2015/03/09/apple_releases_2_lb_12_retina_macbook.html"</A>LilyHay Newman of Slate claims that "you should think of the new macbook asa tablet"</A>. So how true is this? Is the new MacBook like a tablet? <P>Well, no, it's not. The MacBook's screen is nottouch-capable, and is not capable of being used like a tablet screen.The keyboard and touchpad is an integral part of the machine: it isnot optional or detachable. It runs a desktop/laptop operating system(MacOSX), not a tablet operating system such as iOS. The device is nota tablet, it is not "almost a tablet", it is not even likea tablet. It's a small, light, power-efficient laptop. If itmust be compared to something, perhaps it can be compared to a <AHREF="http://www.cs.toronto.edu/~jdd/blog.cgi/2009/02/09#netbooks">netbook</A>,though it has a much better keyboard, touchpad and screen, and is muchmore expensive. <P>Then what about the single I/O port? That's simply the consequence ofthe new <A HREF="http://www.usb.org/developers/docs/usb_31_030215.zip">USB3.1 specification</A>, which finally allows a USB connection to deliverenough power to power a laptop, and defines the USB-C connector, which inaddition to USB data lines, provides "alternate mode" data lines that canbe used for display protocols like DisplayPort. This makes it possible forApple to build <AHREF="http://store.apple.com/us/product/MJ1K2AM/A/usb-c-digital-av-multiport-adapter">multiportadapters for the Macbook</A> that provide video (e.g. HDMI), data (USB-A)and charging ports, making it unnecessary to provide all those portsseparately in the laptop itself. <P>So does this make the Macbook "like a tablet"? While it is true thattablets have been using single connectors for power and data for a longtime, this doesn't make the Macbook tablet-like. It's not the presence ofa single shared power/data connector that makes something like a tablet,it's the interactive screen. Yes, a horse has four legs and is often satupon, but a horse is not anything like a chair. <P> So will I be getting one of the new Macbooks? Probably not: like a finethoroughbred, the new Macbook is lovely but rather too expensive for me.The need to buy the multiport adapter separately makes the already highcost of acquisition even higher. The high price doesn't stop me fromadmiring the design and engineering of this new laptop, but it does keepme from buying one.</description> </item> <item> <title>What's wrong with Blackberry? (and some ideas about how to fix it)</title> <pubDate>Sat, 05 Oct 2013 17:03:00 -0400</pubDate> <link>http://www.cs.toronto.edu/%2525252525252525252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/2013/10/05#blackberry</link> <category>/it</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%2525252525252525252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/it/blackberry</guid> <description>Blackberry is in the news a fair bit these days, and the news seems to be all bad. <AHREF="http://press.blackberry.com/financial/2013/blackberry-reports-second-quarter-fiscal-2014-results.html">Asthe firm reports close to a billion dollars in quarterly losses</a>, a <AHREF="http://www.computerworld.com/s/article/9242767/Update_Gartner_tells_IT_shops_that_it_s_game_over_for_BlackBerry">Gartneranalyst recommends that enterprise customers find alternatives to Blackberryover the next six months</A>. What's the problem?<p>Basically, fewer and fewer people want to buy Blackberry phones. The problemisn't so much that Blackberries don't do what they're supposed to, it'sthat people now perceive iPhones and various Android phones as much betterchoices, and are buying those instead. Why? The reason is that an iPhone oran Android phone isn't the same sort of phone as a traditional Blackberry. AniPhone or an Android phone is a true smartphone, i.e. an "app" phone, aplatform that runs a whole "ecosystem" of third party software. A traditionalBlackberry is a "messaging" phone, a device that specializes in effectivemessaging, such as email. Yes, it can run applications too, but that'snot its primary function, and it shows. <p>To illustrate, consider email. Sending email requires the ability totype quickly. A physical keyboard works best for this, one that stretchesacross the short side of the phone. The screen, located above the keyboard,then becomes roughly square: it can't be very wide, because the phone willthen become too wide to hold easily or to fit in one's pocket, and it can'tbe very tall or the phone will become too long. A square screen is finefor messaging, but for other things that a smartphone might like to do,such as displaying video, one wants a screen that is significantly widerthan it is tall. A smartphone handles this by having a rectangular screen:when doing messaging, one holds the phone vertical: the bottom half of thescreen then turns into a keyboard, and the top half turns into a roughlysquare messaging display. When watching media, such as videos, the phoneis held horizontal, allowing a screen that is wider than it is tall.Hence the smartphone is useful in a broader set of ways: it is not just amessaging device. Smartphones have become good enough at messaging thatmany people do not feel they need a dedicated messaging device. Once thesmartphone is the only device that people feel they need to carry, there'smuch less demand for a messaging phone. <p> Blackberry realized the problem, and tried to create a smartphone of itsown. For instance, in 2008, it released the Blackberry Storm. But itbecame clear that Blackberry's phone OS was not as well suited forgeneral smartphone use as iOS and Android. The Storm was not a commercialsuccess because it did not work as well as competing phones. In response,in 2010 Blackberry bought a company called QNX that had a powerful OS,and started building devices to use it: first the Playbook, released inspring 2011, and then the Z10 phone in early 2013, followed a few monthslater by the Q10 and other phone models. <p>The new Blackberry OS works better than the old in deliveringsmartphone apps, but it was not very mature in 2011, and wasavailable only on a tablet (the Blackberry Playbook). Unfortunately,the Playbook did not sell particularly well because Blackberry <AHREF="http://press.blackberry.com/press/2010/pressrelease-4577.html">badly misrepresented it, calling it the "best professional-grade table inthe industry"</A> though it lacked many features of the market-leading iPad,including key messaging features such as a standalone email client.While it could have been a market success if it were marketed as aBlackberry phone accessory, a role it could effectively play, at releaseit was clearly not a true general-purpose tablet like the iPad. So itaccumulated few apps, while Apple's iOS and Google's Android accumulatedmany. Blackberry realized this fairly quickly, and released an Androidapplication emulation environment for their OS in early 2012, which allowedmany Android apps to be easily moved over to the new OS. But few Androiddevelopers bothered to make Blackberry versions of their Android apps,given the relatively few Playbooks sold. <p> In the meanwhile, Blackberry did itself no favours by making it clearthat there was no future for its existing phones, while failing to delivera phone running its new OS for more than a year. This merely encouragedBlackberry users and app developers alike to switch to another platform.When the Z10 phone finally came out in 2013, the bulk of its apps werethose that had been written for or ported to the Playbook, a far less rich setof applications than any Android or iOS phone. And while the Z10 is a decentphone that comes with some very nice messaging features, Blackberry didnot do an effective job of touting the unique features of the Z10 thatiPhones and Android phones do not have. Moreover, the price was set high(about the same as an iPhone or high end Android phone) and Blackberryproduced a huge number, expecting to sell a great many. Some sold, butmany didn't, and Blackberry's recent $1B loss was due primarily to writingdown the value of unsold Z10s. <p>Blackberry sits today in a difficult position. No, it is not about to goout of business: the company is debt-free and has a couple of billion dollarsin the bank. But its smartphone is not selling. What should it do now? <p>Blackberry's best chance at this point to make its smartphone platformviable is to take its large inventories of written-down Z10 phones andsell them cheaply, using a renewed marketing campaign that focuses on theunique features of the phone's software. The Z10 hardware is really nodifferent than the various Android and iPhone models out there: if thephone is to sell, it has to be on the basis of what makes it unique, andthat's the Blackberry OS software. For instance, Blackberry should showeveryone the clever virtual keyboard that supports fast one-handed typing,the unique messaging hub, and the "Blackberry Balance" software that letsyou separate work items from personal items on the phone. Blackberry needsto hire the best marketing people in the world to help get the message out.This is a "make or break" situation for the platform. <p>Secondly, Blackberry should modify the OS to run Android apps natively,without repackaging. Android app developers are not going to repackagetheir apps for Blackberry. Blackberry needs to recognize this and makesure that Android apps will appear automatically on Blackberry devices.Blackberry will need to find a way to get Google Play (the Android app store)ported to the platform. It is too late to build a separate app ecosystemaround the Blackberry OS: it has to leverage an existing ecosystem, or die.Android is really the only viable option for Blackberry right now. <p>Finally, Blackberry needs to recognize that a niche market for dedicatedmessaging devices exists, and continue making devices that are the bestmessaging phones available, while tapping into an existing app ecosystem.Blackberry needs to be careful not to compromise the devices' effectivenessfor messaging: it should pay attention to how people use the devices in thereal world, and address quickly whatever issues they have. If Blackberrycan't find a way of building such messaging devices using its own OS, itshould switch to Android. Blackberry knows how to make superb messagingphones, and it should find a way to continue to do what it does best.</description> </item> <item> <title>Cloud Computing: Everything Old is New Again</title> <pubDate>Tue, 20 Aug 2013 22:45:00 -0400</pubDate> <link>http://www.cs.toronto.edu/%252525252525252525252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/2013/08/20#cloud</link> <category>/it</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%252525252525252525252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/it/cloud</guid> <description>There is a great deal of hype about Cloud Computing at the moment, andit's getting a great deal ofattention. It's no wonder: when firms such as <AHREF="http://www.bloomberg.com/news/2013-07-21/netflix-rule-as-no-1-s-p-performer-but-.html">Netflix,with a market capitalization of over U$15B</A>, use cloud computing todeliver streaming video services to nearly fortymillion customers around the world, and when the <AHREF="http://www.businessinsider.com/amazon-vs-ibm-in-a-battle-for-the-cloud-2013-7">USCentral Intelligence Agency spends U$600M for cloud computing services</A>,people take notice. But what is it all about? <p> Cloud computing is not really a new thing, it's a variation of a veryold idea, with a new name. In the 1960s, when computers were large andexpensive, not everyone could afford their own. Techniques for sharingcomputers were developed, and firms arose whose business was sellingtime on computers to other firms. This was most commonly described as"timesharing". IBM released its <A HREF="http://www.vm.ibm.com/history/">VMvirtualization environment in 1972</A>, which allowed a mainframe computerto be divided up into virtual computers, each for a different workload.A timesharing vendor could buy and operate an IBM computer, then rent totheir customers "virtual computers" that ran on that machine. From thecustomer's perspective, it was a way to obtain access to computing withoutbuying one's own computer. From the vendor's perspective, it was a way of"renting out" one's investment in computer infrastructure, as a viablebusiness. <p>Today, cloud computing, as did timesharing in the past, involvesthe renting of virtual computers to customers. The name has changed: then,it was called "timesharing"; now, "cloud computing". The type of physicalmachine has changed: then, a mainframe was used to provide computingservices; now, a grid computer. The interconnection has changed: then,leased data lines were typically used; now, the internet. But the basicconcept is the same: a vendor rents virtual computers to customers, whothen use the virtual computers for their computing, rather than buyingtheir own physical computers. <p>The advantages and disadvantages of today's cloud computing echo thepros and cons of yesterday's timesharing. Advantages include risk sharing,the ability to pay for just the amount of computing needed, the option toscale up or down quickly, the option to obtain computing resources withouthaving to develop and maintain expertise in operating and maintainingthose resources, and the ability to gain access to computing resources invery large or very small quantities very quickly and easily. Moreover,cloud computing vendors can develop economies of scale in running physicalcomputers and data centres, economies that they can leverage to decrease thecost of computing for their customers. Disadvantages of cloud computinginclude possibly higher unit costs for resources (for example, clouddata storage and data transfer can be very expensive, especially inlarge quantities), a critical dependance on the cloud computing vendor,variable computing performance, substantial security and privacy issues,greater legal complexity, and so on. These tradeoffs are neithersurprising nor particularly new: in fact, many are typical of "buy"vs. "rent" decisions in general. <p> Then why does cloud computing seem so new? That, I think, is an artifactof history. In the 1970s and early 1980s, computers were expensive andtimesharing was popular. In the 1990s and early 2000s, computers becameincreasingly cheaper, and running one's own became enormously popular.Timesharing faded away as people bought and ran their own computers.Now the pendulum is swinging back, not driven so much by the costof computers themselves, but the costs of datacentres to house them.A few years ago, <A HREF="http://www.amazon.com">Amazon Inc.</A> saw abusiness opportunity in making virtual machines available for rental:it was building grid computers (and datacentres to house them) for itsown operations anyway; why not rent out some of those computing resourcesto other firms? In so doing, Amazon developed an important new line ofbusiness. At the same time, a huge number of new internet firms arose,such as Netflix, whose operations are dominantly or exclusively that ofproviding various computer-related services over the internet, and it madea great deal of sense for such firms to use Amazon's service. After all,when a company's operations are primarily or exclusively serving customerson the internet, why not make use of computing resources that are already onthe internet, rather than build private datacentres (which takes time,money and expertise)? These new internet firms, with lines of businessthat were not even possible a decade or two ago, and Amazon's service,also only a few years old, have lent their sheen of newness to the notionof "cloud computing" itself, making it appear fresh, inventive, novel.But is it? The name is new, yes. But in truth, the concept is almostas old as commercial computing itself: it has merely been reinvented forthe internet. <p> Of course, the computing field, because of its inventiveness, high rateof change and increasing social profile, is rather at risk of falling intotrendiness, and cloud computing certainly has become a significant trend.The danger of trendiness is that some will adopt cloud computing not onits own merits, but solely because it seems to be the latest tech tsunami:they want to ride the wave, not be swamped by it. But cloud computingis complex, with many pros and cons; it is certainly a legitimate choice,as was timesharing before it, but it is not necessarily the best thing foreveryone. It's easier to see this, I think, if we look beyond the name,beyond the trend, and see that the "rent or buy" question for computinghas been with us for decades, and the decision between renting virtualmachines and buying physical ones has often been complex, a balance ofrisks, opportunities, and resources. For an internet firm whose customersare exclusively on the internet, renting one's computing assets on theinternet may make a great deal of sense. For other firms, it may not makesense at all. Deciding which is true for one's own firm takes wisdom andprudence; a healthy dose of historical perspective is unlikely to hurt,and may help cut through the hype.</description> </item> <item> <title>Handling Unsolicited Commercial Email</title> <pubDate>Tue, 23 Apr 2013 12:56:00 -0400</pubDate> <link>http://www.cs.toronto.edu/%25252525252525252525252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/2013/04/23#uce</link> <category>/misc</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%25252525252525252525252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/misc/uce</guid> <description><p> My email address is all over the web: at the time of writing this, <AHREF="http://www.google.ca/search?q=%22jdd%40cs.toronto.edu%22">a searchon google</A> for my email address produces about 15,800 results. Soanyone who wants to find my email address can do so easily. Many people or companies who want to sell me something send me emailout of the blue. I get a great deal of such unsolicited commercialemail, too much to read or pay adequate attention to. I simply deletethem. Unfortunately, many sources of such email persist. So for some timenow, I've elicited the help of technology. I process my incoming email using<A HREF="http://www.procmail.org">procmail</A>, a powerful piece of softwarethat lets me script what happens to my email. When I receive unsolicitedcommercial email, if it is from a vendor or organization I don't have arelationship with, I will often add a procmail rule to discard, unseen, allfuture email messages from that vendor. I've got about 400 organizations(mostly vendors) in my discard list so far, and the list slowly grows.Am I still getting unsolicited commercial email from these sources? I am,but I am not seeing it. It's the same effect, really, as manual deletion(i.e. the message is deleted, unread), but it's easier for me, becauseI am not interrupted. But of course I think it would be better still ifthe email were not sent at all. <p> If you are a vendor with whom I do not have a pre-existing relationship,and you want to send me email introducing your products, please don't.<A HREF="blog.cgi/2011/12/15#coldcalls">I do not accept cold salescallseither.</A> Instead, advertise effectively on the web, so that if I amlooking for a product like yours, I can find you. If you must contactme directly, send me something by postal mail, where, unlike email, thecommunication does not have an interruptive aspect.</description> </item> <item> <title>A closer look at the University of Toronto's international ranking in Computer Science.</title> <pubDate>Thu, 29 Nov 2012 00:00:00 -0500</pubDate> <link>http://www.cs.toronto.edu/%2525252525252525252525252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/2012/11/29#rankings</link> <category>/misc</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%2525252525252525252525252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/misc/rankings</guid> <description><p>International rankings of universities seem to be all the ragethese days. The interest seems to be fed by three rankings ofparticular prominence that have emerged in the past decade. Theseare <A HREF="http://en.sjtu.edu.cn">Shanghai Jiao Tong University</A>'s <A HREF="http://www.shanghairanking.com/">Academic Ranking of World Universities</A> (sometimes knownas AWRU, or simply as the "Shanghai Ranking"), <A HREF="http://www.qs.com">QuacquarelliSymonds</A>' <A HREF="http://www.topuniversities.com">QSWorld University Rankings</A>, and the <AHREF="http://www.timeshighereducation.co.uk/world-university-rankings/">TimesHigher Education World University Rankings</A>. Part of the attractivenessof these rankings is that they can become a way of "keeping score", ofseeing how one institution does in comparison to others. <p>My employer, the <A HREF="http://www.utoronto.ca">University ofToronto</A>, does quite well in these rankings, particularly my department,<A HREF="http://www.cs.toronto.edu">Computer Science</A>. The subjectarea of Computer Science is not ranked separately in the <A HREF="">TimesHigher Education World University Rankings</A> (it's bundled together withEngineering), but in the other two, Toronto has consistently ranked in thetop ten in the world each year in Computer Science, with only one exception. <p>This exception is recent, however, and worth a closer look. In the <AHREF="http://www.topuniversities.com">QS World University Rankings</A>for Computer Science and Information Systems, Toronto dropped from <AHREF="http://www.topuniversities.com/university-rankings/world-university-rankings/2011/subject-rankings/technology/computer-science-information-systems">10thin 2011</A> to <AHREF="http://www.topuniversities.com/university-rankings/world-university-rankings/2012/subject-rankings/technology/computer-science-information-systems">15thin 2012</A>. This big drop immediately raises all sorts of questions: hasthe quality of Toronto's Computer Science programme suddenly plummetted?Has the quality of Computer Science programmes at other universities suddenlysoared? Or has the QS World University Rankings changed its methodology? <p>To answer this question, let's look at how other universities havechanged from 2011 to 2012 on this ranking. Many (MIT, Stanford, Berkeley,Harvard, Oxford, Cornell, and others) stayed where they were. Others droppedprecipitously: Cambridge University dropped from 3rd to 7th, UCLA from 8th to12th, and Caltech plummetted from 7th to 27th. Some other universitieswent up: Carnegie Mellon University (CMU) went from 9th to 3rd, ETH Zurichfrom 11th to 8th, the National University of Singapore (NUS) from 12thto 9th, and the Hong Kong University of Science and Technology (HKUST)soared from 26th to 13th. Surely these curious and significant changesreflect a methodology change? But what? <p>The QS university rankings website, in the Methodology section, Academic subsection, reveals <A HREF="http://www.iu.qs.com/university-rankings/subject-tables/?">something of interest:</A><PRE> NEW FOR 2012 - Direct Subject Responses Until 2010, the survey could only infer specific opinion on subject strength by aggregating the broad faculty area opinions of academics from a specific discipline. From the 2011 survey additional questions have been asked to gather specific opinion in the respondent's own narrow field of expertise. These responses are given a greater emphasis from 2012.</PRE>To understand this change, it needs to be recognized that the QS rankingsrely highly on the opinions of academics.A large number of academics around the world aresurveyed: the QS rankings website indicates that in 2012, <AHREF="http://www.iu.qs.com/university-rankings/academic-survey-responses/">46079academic responses were received, of which 7.5% addressed ComputerScience." </A> The seemingly modest change made in 2012, to weigh moreheavily the opinions of academics in a field about their own field, givenits impact on the 2012 results for Computer Science, leads one to wonderabout the regional distribution of academics in Computer Science incomparison to academics in other disciplines. One significant factor maybe China. <p><AHREF="http://fusion.grids.cn/career/attachments/china-SIGCSE2010.pdf">In1999, courses in the fundamentals of computer science became requiredin most Chinese universities, and by the end of 2007, China had nearly amillion undergraduates studying Computer Science</A>. While QS rankingsdoes not indicate regional distribution by discipline for the academicswhose opinions it consults, the surge in the number of Chinese computerscientists worldwide in the past decade almost certainly must have aneffect on the regional distribution of academics in Computer Scienceas compared to other disciplines. As such, is it any surprise to seeworld universities prominent in China that possess strong Computer Scienceprogrammes (such as HKUST and NUS) climb significantly in the rankings,and others less prominent in China plummet? But if a world ranking ofuniversities is so affected by regional shifts in those whose opinionis being solicited, how reliable is it as an objective gage of the realquality of a given university? <P>Perhaps a more reliable gage of quality can be foundin the Shanghai ranking, which is not opinion-based, but <AHREF="http://www.shanghairanking.com/aboutarwu.html">relies on concreteindicators and metrics</A>. On the Shanghai ranking, the University ofToronto consistently ranks 10th in the world in Computer Science in 2010,2011, and 2012. But what does this mean, concretely? <P>To answer these questions, we need to grapple with animportant fact: in Computer Science, the US dominates. As a nation, <AHREF="http://books.google.ca/books?id=jv1m50j4ZAMC">theUS has been enormously supportive of Computer Science ever since the fieldfirst existed, and as a result, it has become pre-eminent in computing</A>.Nine of the top ten schools in the Shanghai ranking, and twenty of thetop twenty-five, are in the US. For the University of Toronto tobe one of the handful of universities outside the US to break intothe top twenty-five, and the only one to break into the top ten,is a significant accomplishment. A chart is illustrative: <p><IMGSRC="http://www.cs.toronto.edu/~jdd/blog/misc/rankings/shanghai-cs-nonus-2012.jpg" width=800 height=600> <p>Of course, the University of Toronto is in Canada, so a comparisonto other schools in Canada is also illustrative. For Computer Science,on the Shanghai ranking, there seems to be no close Canadian rival.In 2012, UBC comes closest, being a only a few points short of breakinginto the top 25, but all other Canadian schools rank well back:<p><IMG SRC="http://www.cs.toronto.edu/~jdd/blog/misc/rankings/shanghai-cs-can-2012.jpg" width=800 height=600> <P>Even compared to other disciplinesthat have Shanghai rankings (only science, social science, and relateddisciplines seem to be ranked), Toronto's pre-eminence in Computer Sciencein Canada is striking:<p><IMG SRC="http://www.cs.toronto.edu/~jdd/blog/misc/rankings/shanghai-can-2012.jpg" width=800 height=600> <p>From a score-keeping perspective, I think we can conclude that theUniversity of Toronto is doing very well in Computer Science with respectto other universities in Canada, and it is one of the few non-USschools that can keep up with the US in this field. <p>But all this needs to be put into perspective. After all, rankingsare not a full picture, they're aggregations of metricsof varying value, they represent a formulaic approach to something(university education) that cannot always be so conveniently summarized,and they reflect methodologies chosen by the producers of the rankings,methodologies that may not always best reflect objective quality.Of course, if the University of Toronto were to climb to fifth, I'd bepleased, and if it were to drop to fifteenth, I'd be disappointed: surelythe score-keeper in me can be allowed this much. But in the overall schemeof things, what matters most for Computer Science at Toronto is not ourscore on a ranking system, but the objective quality of our programme,the learning outcomes of our students, and the impact of our research, andthese things, not our score on rankings, must always remain our top priorities.</description> </item> <item> <title>Intel desktop CPU price-performance: Hyperthreading not helping?</title> <pubDate>Wed, 22 Aug 2012 14:07:00 -0400</pubDate> <link>http://www.cs.toronto.edu/%252525252525252525252525252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/2012/08/22#intelcpus</link> <category>/it</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%252525252525252525252525252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/it/intelcpus</guid> <description>Typically, CPU prices follow performance. Faster CPUs command higherprices; slower CPUs are available for less. Recent Intel desktop CPUs continue to show this general pattern, but there appears to be more to the story than usual. <P> At first glance, everything seems to be what you would expect. Using currentpricing in US$ at time of writing from <A HREF="http://www.newegg.com">newegg.com</A>, we get:<table border=1><tr><th>Processor</th> <th>PassMark</th> <th>Price</th> <th>PassMark/$</th> <th>Price-Performance vs G640</th></tr><tr><td>Pentium G640</td> <td>2893</td> <td>$79</td> <td>36.6</td> <td> 100%</td></tr><tr><td>i3-2120</td> <td>4222</td> <td>$125</td> <td>33.8</td> <td> 92.2%</td></tr><tr><td>i5-3570</td> <td>7684</td> <td>$215</td> <td>35.7</td> <td> 97.6%</td></tr><tr><td>i7-3770</td> <td>10359</td> <td>$310</td> <td>33.4</td> <td> 91.3%</td></tr></table>The <A HREF="http://www.cpubenchmark.net/">PassMark (http://www.cpubenchmark.net/</A>) todollar ratio is pretty consistent across all these processors, roughly 35 &plusmn; 2. <P> But what happens if we look at a more real-life benchmark? ConsiderSPEC CPU 2006 Integer (CINT2006) Baseline. For each CPU, I used the CINT2006 Baseline results fromthe most recently reported Intel reference system, as reported on <A HREF="http://spec.org">spec.org</A>. In the case of the G640, noIntel reference system was reported, so I used the results fora Fujitsu Primergy TX140 S1p.<table border=1><tr><th>Processor</th> <th>CINT2006 Base</th> <th>Price</th> <th>CINT/$</th> <th>Price-Performance vs G640</th></tr><tr><td>Pentium G640</td> <td>34.4</td> <td>$79</td> <td>0.44</td> <td> 100%</td></tr><tr><td>i3-2120</td> <td>36.9</td> <td>$125</td> <td>0.30</td> <td> 67.8%</td></tr><tr><td>i5-3570</td> <td>48.5</td> <td>$215</td> <td>0.23</td> <td> 51.8%</td></tr><tr><td>i7-3770</td> <td>50.5</td> <td>$310</td> <td>0.16</td> <td> 37.4%</td></tr></table>When looking at CINT2006 Baseline, we see the price-performance ratio drop off dramatically asthe processor price increases. We would expect this from the i3 to the i5,since SPEC cpu int is a single job benchmark and the i3 to the i5 representsa transition from two to four cores, but it's curious to see the dropoffin the price-performance ratio between the G640 and the i3 (both dual-coreCPUs), and the i5 and the i7 (both quad-core CPUs). What might be going on? <P> A look at hyperthreading may provide some answers. Intel hyperthreading is a featureof some Intel CPUs that allow each physical core to represent itself tothe OS as two different "cores". If those two "cores" simultaneously runcode that happens to use different parts of the physical core, they canproceed in parallel. If not, one of the "cores" will block. The i3 andi7 CPUs offer hyperthreading, the Pentium G and i5 do not. It turns outthat the PassMark benchmark sees significant speedups when hyperthreadingis turned on. SPEC CINT2006, and many ordinary applications, do not. <P> What about SPEC CINT2006 Rate Baseline, then? The SPEC CPU Rate benchmarks measurethroughput, not just single-job performance, so maybe hyperthreading helpsmore here? Let's see:<table border=1><tr><th>Processor</th> <th>CINT2006 Rate Base</th> <th>Price</th> <th>Rate Base/$</th><th>Price-Performance vs G640</th></tr><tr><td>Pentium G640</td> <td>61.7</td> <td>$79</td> <td>0.78</td> <td> 100%</td> </tr><tr><td>i3-2120</td> <td>78.8</td> <td>$125</td> <td>0.63</td> <td> 80.7%</td> </tr><tr><td>i5-3570</td> <td>146</td> <td>$215</td> <td>0.68</td> <td> 87.0%</td> </tr><tr><td>i7-3770</td> <td>177</td> <td>$310</td> <td>0.57</td> <td> 73.1%</td> </tr></table>If we look at the transition from two to four cores (by comparing the i3 tothe i5), we now see that the price-performance of the i5 is better than the i3:this is no surprise, since we are now measuring throughput, and from the i3to the i5, we go from two to four cores. But there still is a dropoff inprice-performance between the Pentium G and the i3, and again between thei5 and the i7. It's not as extreme as before, but it is still significant.This suggests that hyperthreading may help with throughput, but not as muchas the increase in price would suggest. <P> What does this mean, then? It suggests the increase in price from anon-hyperthreaded to a hyperthreaded Intel desktop processor may reflect morean increase in PassMark performance than an increase in real performance.Hyperthreading may have a positive effect, it seems, but typically not asmuch as PassMark suggests. At present, for best real-world price-performancein Intel desktop CPUs, I would consider models without hyperthreading.</description> </item> <item> <title>How to avoid being fooled by "phishing" email.</title> <pubDate>Tue, 26 Jun 2012 16:56:00 -0400</pubDate> <link>http://www.cs.toronto.edu/%25252525252525252525252525252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/2012/06/26#phishing</link> <category>/it</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%25252525252525252525252525252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/it/phishing</guid> <description>A "phishing" email is an email message that tries to convince you toreveal your passwords or other personal details. Most often, it triesto send you to a website that looks like the real thing (e.g. your bank oryour email provider) but is really a clever duplicate of the real websitethat's set up by crooks to steal your information. Often the pretence looksauthentic. If you fall for it and give your password or other personaldetails, criminals may steal your identity, clean out your bank account,send junk email from your email account, use your online trading accountto buy some penny stock you never heard of, send email to all the peoplein your address book telling them you're stranded in a foreign country andneed them to wire money immediately, or do any number of other bad things. <p> But there's a really easy way to avoid being fooled by phishing messages.If you get a message that asks you to confirm or update your account details,<em>never, ever go to the website using a link that is in the email messageitself.</em> Remember, anyone can send you a message with any sort offraudulent claim, containing any number of links that pretend to go to oneplace, but really go to another. So if you feel you must check, go to thewebsite that you know for sure is the real thing: use your own bookmark(or type in the URL yourself), not the link in the message. </description> </item> <item> <title>Dealing with unsolicited salescalls (cold calls).</title> <pubDate>Thu, 15 Dec 2011 15:14:00 -0500</pubDate> <link>http://www.cs.toronto.edu/%2525252525252525252525252525252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/2011/12/15#coldcalls</link> <category>/misc</category> <guid isPermaLink="false">http://www.cs.toronto.edu/%2525252525252525252525252525252525252525252525252525252525252525252525252525257Ejdd/blog.cgi/misc/coldcalls</guid> <description><p> For many years, I've been plagued by unsolicited salescalls. It'snot very hard to find my phone number, and various people (mostly inthe IT realm) call me up out of the blue hoping to sell me something.The interruption is unwelcome, even if the product isn't. <p> For some years now, my policy is to explain to the caller that I don'taccept unsolicited salescalls, sincerely apologize, and end the call.Occasionally, I am then asked how I am to be contacted. I explain that Iprefer to do the contacting myself: when I have a need, I am not too shyto contact likely vendors and make inquiries about their products. <p> Occasionally I run into someone who is offended by my unwillingness totake their unsolicited salescall. I do feel more than a little sympathyfor the salesperson when this happens: I imagine they may think I objectedto something they did, or to their manner. The fact is, I handle allunsolicited salescalls this way. As for whether it is intrinsiclyoffensive to reject unsolicited salescalls out of hand, I don't thinkit is. Indeed, it is natural for a salesperson to want their salescall,even if unsolicited, to be better accepted. But it is unreasonable forany salesperson to expect that unsolicited sales inquiries to strangerswill always be welcome. But I do apologize, each time, and in general,when I so quickly end telephone conversations with salespersons who callme out of the blue. <p>Dear reader, if you are a salesperson, and you are tempted to contactme to sell me something, please do not call. Instead, just advertisegenerally (and if you must, send me some mail in the post). Trust me tofind you when the need arises. I frequently do.</description> </item> </channel></rss> If you would like to create a banner that links to this page (i.e. this validation result), do the following:
Download the "valid RSS" banner.
Upload the image to your own server. (This step is important. Please do not link directly to the image on this server.)
Add this HTML to your page (change the image src attribute if necessary):
If you would like to create a text link instead, here is the URL you can use:
http://www.feedvalidator.org/check.cgi?url=http%3A//www.cs.toronto.edu/%257Ejdd/blog.cgi/index.rss