Latest News

Using this technology, students from different locations can play music together as if they were in the same room. (cim.mcgill.ca)
a, Science & Technology

Shared reality enriches human-computer interaction

Imagine arriving late to class, ears burning from the cold, and stepping over puddles and bags until you finally find a seat. As the professor lectures away, a student several rows down raises his hand. The professor, unaware, continues the lecture. After some time, the student eventually puts it down.

For most McGill students, this scenario isn’t hard to imagine. Professors might not always be fully aware of the hands that go up at the top or the far sides of a lecture hall, and it’s not their fault. They’re only human after all.

What if there was a way to detect if a student raised their hand in a large crowd, or to recognize if students were confused during a lecture? These possibilities are being explored at the McGill Shared Reality lab.

“A prototype system that we developed years back was used to automatically detect if a student raised his or her hand up for a question,” Dr. Jeremy Cooperstock, Director of the Shared Reality Lab, said. “Since it was in a large class setting, the system would then inform the instructor by raising a flag on the screen about who has their hand up, and what part of the room they’re in.”

Experiments at the Shared Reality Lab involve the use of virtualized reality techniques and advanced networking protocols to give users a strong sense of co-presence—the feeling of being together in a shared physical environment. This is accomplished using a number of screens, cameras, projectors, and microphones, along with a high-fidelity vibrosensory system.

“This high fidelity and low latency type of communication gives people the ability to feel like they are engaged in coordinated synchronous activity with those far away, while in the comfort of McGill,” Cooperstock explained.

The lab was one of the first research groups in the world to support a distributed music performance. Jazz students from Montreal and Stanford performed together at the same time using the technology developed by the Shared Reality Lab. They were able to see and hear each other in the same capacity as if they were physically in the same space.

cim.mcgill.ca
cim.mcgill.ca

The lab also looks at how to adapt users to a different kind of environment by rendering physical scenarios, such as the sensation of walking on different ground surfaces, such as snow, gravel, or sand.

“And they can experience that even though they are physically in a laboratory environment, walking on floor tiles,” Cooperstock explained.

The third dimension of the lab involves looking at sensory substitutions for those unable to experience a certain aspect of the everyday world around them. By working with the blind community, the Shared Reality Lab looks for ways to give these people the visual experience of the world around them. This is accomplished by providing the information usually available to vision through audio, explains Cooperstock. The lab has also given demonstrations on new technologies for the Android and iPhone, which give users a constant display, through audio, of what points of interest are around them while walking outside.

Along with developing sensory substitutions, the lab has created different applications to help Music and Medical students at McGill. For musical training, the lab developed a simulator known as “Open Orchestra” which has received significant recognition.

“The simulator was [developed] to give classical and jazz musicians the experience of rehearsing with the rest of their band or the orchestra around them,”  Cooperstock explained. “It wasn’t a live performance scenario. Rather, it was a rich multimodal experience of what it feels like, looks like, and sounds like to be sitting in, for example, the second violinist seat in a 30 seat orchestra and playing along with the different musicians, while seeing and hearing the conductor at the same time.”

Similarly, a training system was used to train McGill medical students in response scenarios. This project was funded by the Canarie’s Network Enabled Platforms (NEP) program, and was completed in 2010. It involved using medical mannequins to mimic different physiological functions, such as blood flow to the heart, and teaching students through scenarios where they had to experience and address real situations, such as a patient in a car accident.

Through this new approach to human-computer interactions, the Shared Reality Lab offers a glimpse into the future of virtualized reality, which could change the way we interact with people around the world. One day, attending the opera could involve listening to different opera singers in different time zones, standing on different stages in a worldwide opera house located in cyberspace.

conncad.com
a, Science & Technology

Behind the scenes at the Dent lab

The Dent Lab in the Stewart Biology building is humming with activity. Run by Dr. Joseph Dent, an associate professor and researcher at McGill University, the lab focuses on the molecular genetics of the behaviour in C. elegans, a nematode roundworm. 

Specifically, the lab’s research focuses on understanding the structure and function of neurotransmitter receptors, the role they play in behaviour, and how we can manipulate them to treat diseases, or better understand how nervous systems work. Essentially, neurotransmitter receptors are membrane receptors that receive electrical signals, facilitating the transmission of information from the brain to the body, and vice versa.

“Our lab has basically two components,” explains Dent. “One is a relatively applied component, and the other a more basic research component.”

The applied component of the lab concerns the relationship between neurotransmitter receptors—important targets for antiparasitic drugs—and pesticides. Dent and his team are currently looking into how existing drugs kill parasites, specifically nematode roundworm parasites, and how nematode parasites develop mutations that allow them to become resistant to these drugs. They hope, through this research, to learn how we can use antiparasitic drugs to better prevent the disease from reoccurring, as well as to make the drugs more effective against resistance developed by parasites.

Nematode roundworm parasites are of significant importance due to the disease caused by the nematode Onchocerca volvulus. River blindness, caused by O. volvulus, is endemic to Sub-Saharan Africa, where 18 million people are at risk of losing their sight. The disease is currently being treated with the drug Ivermectin, which is given in yearly doses by the World Health Organization (WHO) to help people who are already affected and reduce the rate of transmission.

This second area of research comprises the lab’s more basic research aspect. The team has investigated the role neurotransmitters play in behaviour, how the nervous system uses them to modulate behaviour in interesting ways, and the fundamental features of the neurotransmitters themselves. Through this work, they are able to look at the mechanisms behind Ivermectin resistance and how nematodes develop resistance to the drug.

The team works with the roundworm C. elegans in its experiments. While not parasitic, this organism is a much more efficient model to use during experimentation. Since C. elegans is highly similar to other organisms, the team can translate what they learn to various other systems.

“It turns out nematodes have a lot of receptors that humans do not have. These are a good target for anti-parasitic drugs. You want a drug that targets the parasite and not humans,” says Dent.

The Dent Lab works with C. elegans first, and then collaborates with the Institute of Parasitology in order to transfer the work performed on C. elegans to see if it has a similar effect on the parasite.

Looking to the future, Dent says, “We’d love to come up with a new, effective, safe drug that would allow us to have an impact on these diseases.”

“We would also like to better understand the design of the neurotransmitters in all organisms, [in order to] use the information to better focus or target our search for drugs to specific subsets of channels,” explains Dent. “If we understand it, can we design better drugs and better drug targets that are less likely to develop resistance? If we understand how resistance occurs before it occurs, can we use the drugs in better ways?”

Behind the scenes of the lab’s operations, funding plays an important role. According to Dent, the Lab does not get as much support as he would like, since the Canadian health agencies are focused on research concerned more directly with Canadian health. However, the lab receives more support from the agricultural and pharmaceutical industry, which is interested in these drugs because they can also be used to treat livestock. Ivermectin, for instance, is an active ingredient in drugs used by farmers to treat livestock with deworming agents.

“The economics are such that the companies make their money selling these drugs to farmers to treat livestock, so that they can produce cheap meat. We benefit from that, in the sense that we receive industry money to support this research.”

Most important, however, is the idea that spurred this research. When asked, Dent explains it was “just lucky.” While studying the eating behaviour of C. elegans as a post doctoral fellow, Dent discovered a mutant gene that affected their behaviour, and made it less efficient. He located and cloned the gene, only to discover it was a neurotransmitter receptor and the target of the drug Ivermectin—from there, his research took off.

“It often happens that you study one thing and make a completely unexpected observation. What is exciting about research is that, you never know where it will go next—the most exciting research is the research you didn’t anticipate on doing.”

 

Whitney Pang (Alexandra Allaire / McGill Tribune)
a, Opinion

Why can’t boys like cartoon ponies?

The season three finale of Hasbro’s animated series My Little Pony: Friendship is Magic aired earlier this month.  The children’s show has attracted an unexpected adult male fan base called “bronies,” a phenomenon that has garnered attention and sparked much gender-based debate.  A long history of gender-exclusive children’s programming has formed gender archetypes in viewership. From an early age, we are conditioned to have fixed social expectations of what plot content and character traits boys and girls should or should not like.  I find this problematic. It deters viewers from adopting seemingly benign, yet often very important, traits generically attributed to the opposite sex, and hinders any social progression in redefining gender archetypes.

Gender-exclusive programming as a strategy of market segmentation has long been a dominant pattern in the children’s entertainment industry. Producers identify and isolate a particular gender and age group, and tailor the programming preferences of each market to guarantee viewership. Producers then secure the attention of audiences to create the “audience commodity,” which is sold to advertisers.

According to the 2001 study “Development of Gender Differences in Children’s responses to Animated Entertainment” from the journal Sex Roles, among children ages three to five, boys showed a much greater preference for violent content, while girls were rather more interested in romantic drama.  Consequently, programs marketed specifically at boys, such as Transformers or Ben 10, primarily feature male characters possessing masculine attributes of leadership, bravery, and aggression. Programs for girls contrarily mainly portray female protagonists and their struggles with relationships. Friendship is Magic, for instance, follows the adventures of six female ponies as they discover the importance of friendship. When chidren’s programs aimed at specific genders repeat gender codes, the implicit message is: “This is what you are supposed to like.” This influences the gendered preferences of young viewers, which in turn determines content of future shows.  There is a vicious cycle of constant reiteration of gender archetypes, one that allows little space for deviation.

[pullquote]When chidren’s programs aimed at specific genders repeat gender codes, the implicit message is ‘This is what you are supposed to like.[/pullquote]

So what happens when these gender archetypes are challenged? The emergence of “brony” subculture has demonstrated the rigidity of the public’s expectations of gender performativity. Many commentators have expressed their disbelief that a show for little girls could be so popular among males, seeing this trend as unexpected, or even intolerable.  Fox News commentator Andy Levy stated that these men “should feel shame” for their obsession. “Bronies” have radically rewritten expectations of gender behaviour by appreciating a show that embodies benign—though supposedly feminine—virtues of love and friendship.  I would argue that this is a form of social progress, even while expectations of gender behaviour have deemed the “brony” subculture as something highly unnatural.

As ingrained as gender archetypes are, the public should embrace the these roles as social progress. Perhaps to redefine viewer archetypes in a more subtle way (as opposed to the radical nature of “brony” fandom), producers should further explore the possibilities of gender-neutral programming: shows that seek to portray both male and female protagonists possessing atypical gender codes.  Non-gender exclusive shows such as Annie Award nominees Avatar: The Legend of Korra and Adventure Time have recently found immense popularity among both boys and girls. A new era of gender-neutral programming may be on the horizon.

 

a, Opinion

A role for chance?

Everyone who gets into an elite university, including our own, probably feels at one time or another that they were deserving of meriting acceptance into the school, and that the admissions committee did not reach this decision lightly.  A highly revealing article published by Ron Unz in The American Conservative entitled “The Myth of American Meritocracy” calls this assumption into question. In this piece, Unz raises uncomfortable questions by parsing vast amounts of data to show that certain racial and ethnic groups are disadvantaged by standard admission practices (at least in Ivy League schools), and generally calls into question the utility of an admissions committee at all.

While focusing on how many ethnic groups are disadvantaged by admission committees, Unz’s clearest example of discriminatory practices are those that relate to Asian-Americans. He points to a troubling fact that since 1993, Ivy League acceptance of Asian-Americans has declined from 20 per cent to 17 per cent, despite a growing population and increases in household income. Though they comprise only four to five per cent of the U.S. population, Unz notes that Asian-Americans make up about 30 per cent of high school National Merit Scholars, and form the majority of participants in high-level math and physics competitions.

Statistically speaking, Asian-Americans must score 140 points higher on their SATs to be given the same standing as white counterparts. All else equal, it seems that Asian Americans should form a larger proportion of the student population than they currently do. So what is leading to this impasse?

[pullquote]By indulging in chance and eliminating bias, universities will see many salutatory benefits.[/pullquote]

It is obvious that if some groups are being under-represented, then there are other groups who are the exact opposite. While not assigning any nefarious motives, Unz points out that the American Jewish population is likely over-populated in the Ivy Leagues. This is the case for a combination of reasons, he concludes, including Jewish academic merit, prior guilt from excluding Jews from Ivy Leagues, and the prevalence of Jews in high administrative positions in elite universities. Unz speculates that this mix of factors has allowed subtle biases to influence the admissions process, stymieing the chances for other groups to be admitted into these schools.

Unz’s solution is simple. Admit certain clearly deserving students to fill a small portion of your class, such as those who have won various academic awards or achieved a perfect SAT score. For the more difficult task of filling in the majority of an incoming class, Unz suggests setting a minimum threshold of acceptability, and then picking students through a lottery system. By indulging in chance and eliminating bias, universities will see many salutatory benefits. While the selection process would not be a perfect meritocracy, it would bring the schools closer to reaching this stated goal. Secondly, Unz hypothesizes that students would feel more humble about getting into these elite schools, knowing that their acceptance was predicated on chance, and laying waste to the notion that getting into a top school is completely merit-based.

Luckily, many of these afflictions found in the United States do not plague McGill, at least not to the same extent. In Fall 2009, McGill admitted 48 per cent of students, a rate much higher than the Ivy Leagues, and with less selectivity comes a lesser chance of applying under-the-table admissions practices. However, Unz’s idea to introduce lottery based admissions schemes could have a great benefit. First, this idea could save the university money by eliminating some of its admissions committee. McGill administration staff, which includes its admissions committee, made approximately $22 million in salary, according to the 2011 McGill budget. A computer is a lot cheaper than a staff of workers, and—if Unz is right—possibly more effective. Second, a lottery would eliminate any chance of overt or hidden biases affecting the chances of a student getting into McGill. In a field where picking students to attend your university is more of an art than a science, maybe it’s time to lay our fates in the hands of chance.

Genome sequencing helps predict the inheritance of diseases (lejeuneusa.org)
a, Science & Technology

DNA testing offers healthier future

Just before I was born, my parents consulted an astrologist to find out if I would be born healthy. Using the stars, the astrologist made predictions about my mental and physical development. 

Before my kids are born, I might go to a company such as 23andMe. 23andMe uses DNA sequencing to recognize if babies will be genetically susceptible to certain diseases as they grow older.

The technology used by 23andMe was made possible due to the Human Genome Project, which successfully sequenced the complete human genome. Published in 2003, the human genome was to be the harbinger of the age of personalized medicine. However, the costs of whole genome sequencing were prohibitively expensive, and time consuming.

Costs have dropped dramatically in the past few years, increasing accessibility of sequencing the genome to a large number of people. The National Human Genome Research Institute (NHGRI), which tracks the costs of sequencing a genome, estimates current prices to be around $7000, down from around $50,000 when the human genome was published, due to improvements in technology.

The time to sequence these genomes has been reduced as well. In October 2012, a research group from Children’s Mercy Hospital, in Kansas City, Missouri, managed to sequence the whole genome of a baby in 50 hours.

Currently, there are a number of companies that offer to test single genes or a panel of genes related to a certain disease. While useful, these tests, which range between $500-1,000, might only be prescribed in case the family has a history of that disease.

Whole genome sequencing offers the ability to analyze the entire genome, thereby providing clues to what diseases may arise, irrespective of heredity. Additionally, genome sequences can help provide evidence of extremely rare hereditary diseases, which the targeted genetic testing provided by companies may not reveal.

Finally, knowing the genome of the newborn child or fetus could help provide appropriate genetic counselling to parents early on, and help physicians design targeted therapies. The decrease in costs, as well as the benefits whole genome sequencing provide, has dramatically increased demand for this procedure from the public. Soon, it may be the norm to obtain a whole genome sequence of a baby as soon as it is born, and use this data to determine future medical intervention.

Whole genome sequencing, however, is not without its drawbacks. The New York Times, in a piece on the procedure, states that out of 100 patients with any genetic disorder that was sequenced, only 30 patients had a misprinted gene detected by whole genome sequencing. Of these, only three per cent of the patients receive better management, and only one per cent gets treatment.

As of now, whole genome sequencing to identify diseases is still like looking for a needle in a haystack. We do not know all the mutations in a gene that could give rise to a disease. Most diseases arise from defects in multiple genes, further complicating diagnosis. Even if genes suggesting increased risk of disease were to be identified, we lack the ability to treat many of these diseases with current medication.

Improvements in genome sequencing and reduction in costs will someday help doctors better tailor medicine towards the individual patient. With developments in DNA testing, parents will no longer have to consult an astrologist to predict their kids’ physical and mental development. Instead, a DNA sequence will provide insight into their susceptibility to disease. Scientifically, this is the way forward.

 

a, Science & Technology

How does our memory work?

The human brain, composed of over 100 billion cells, is a natural work of art. Groups of brain cells, called neurons, and their synapses—the gaps in between them—are the functional units of the brain that allow us to store memories. While these cells are responsible for what we remember, what enables us to recall our sixth birthday party, but forget what the professor talked about 15 minutes ago in class? Welcome to the mysterious world of the human memory.

The difference between remembering that birthday party over the lecture material revolves around the differences in how we organize and process short and long-term memories. Psychologists view the memory system as analogous to a computer in terms of consisting of three phases: encoding, the process of transforming information into a neural code that the brain can understand; storage, the retention of that code; and retrieval, the act of recalling that information.

The modal memory model proposed by Richard Atkinson and Richard Shriffin in 1968 gives a deeper understanding of how memory works, by breaking it down into three types of memory. Sensory memories are the most brief form of memory, lasting for a fraction of a second. An example of sensory memory is when you look at something and quickly glance away. Due to your sensory memories, you can recall only some of the object’s details. It allows us to take in the world as a continuous stream, rather than discrete chunks of information.

By contrast, short-term memory, also known as working memory, holds data in awareness for a brief period of time, during which our brain can use it to carry out several necessary processes. The retention period for short-term memory is around 20 to 30 seconds. For instance, when looking up a telephone number and walking across the room to dial it, your working memory is at work. Unless you repeat this task over and over again, some digits begin to fade by the time the actual number is dialed.

Another aspect of short-term memory is memory span, which refers to the number of distinct items that can be stored in the working memory at a time. Initially, this number was suggested to be seven (plus or minus two) by George Miller. However, a more recent study conducted in 2005 claims that the span may be as little as four items. Next time you try to remember a phone number for a friend, don’t feel bad when you forget a digit—your short-term memory really only retains four at a time.

On the other hand, long-term memory, the more permanent component of the memory system, is a much less transitory system. Essentially, long-term memory logs new information by associating it with other related topics already stored in your memory.

This is why it is harder to learn something completely new, as opposed to learning about something you’re already familiar with. It’s also why you shouldn’t wait to take Calculus 3 long after you have completed Calculus 2.

Furthermore, the precision with which your neurons store information depends on the strength of the neural associations formed in your brain. It is much easier to remember facts related to information already stored in your memory than to learn a wholly novel concept.

But not all content of the long-term memory is reliable. McGill professor Karim Nader proposed the process of reconsolidation, which explains why some of our long-term memories are distorted. According to Nader, once memories are activated or retrieved, they must be consolidated to be stored in the brain as memory. An improper restoration could ultimately lead to distortions in your memory. This concept could be used to treat trauma patients by helping them retrieve painful memories, and then disrupting the reconsolidation process to alleviate the pain associated with those recollections.

a, Science & Technology

When Mendel meets Darwin

First there was Darwin, who published the revolutionary On the Origin of Species in 1859, introducing the world to the theory of natural selection. According to this theory, genetic variations arise as adaptions to differential environments, where any profitable traits that incur a survival advantage are selected for and preserved through inheritance. In other words, if a long nose gave you the upper hand over other members of your species, the population would evolve to favour large snouts.

creationrevolution.com
creationrevolution.com

Enter Gregor Mendel and the re-discovery of his pea experiment. Mendel cross-bred different combinations of peas in order to illustrate patterned inheritance of genes from one generation to the next. A contemporary of Darwin, the scientific community did not accept the significance of Mendel’s work until the 20th century, when DNA was identified as the molecule that was passed down from parents to their offspring.

After this recognition, Mendelian genetics could be combined with natural selection, placing a new emphasis on reproductive success. In the modern evolutionary view of natural selection, it is imperative that an organism survives until the age of sexual maturity in order to pass on their genes. ‘Mutation proposes and nature disposes,’ with regards to new genetic traits.

Interestingly, throughout this evolutionary discourse, it was assumed that natural selection exerts its pressures only on the organism. But what about our cells? Our bodies are made up of trillions of cells—would they not be exposed to the same selective pressures?

Scientists asked these questions too, adopting an accepted model of cellular and genetic integrity. According to this model, while there are different types of cells with different functions, each cell contains the same genetic profile. Whether DNA is extracted from our blood or our hair, the sequence should theoretically be the same—the only notable exception is the distinctive mutations observed in cancer cells.

Recently, with the advancements made in biotechnology, this model came into question. For the first time in history, we have the tools to readily sequence different tissues at will. As a result, some evidence shows that we are not composed of a single genetic sequence, but rather, a mosaic of many genetic sequences specific to different tissues.

How does this genetic diversity arise? McGill professor of the Department of Human Genetics Dr. Bruce Gottlieb explains:

“Studies have shown that early on in fetal development, you can get the DNA repair mechanism turned off in certain tissue, and you acquire variant [gene sequences].”

This variation leads to a kind of cellular ‘survival of the fittest,’ where the cells with the evolutionary upper hand prosper, compared to others who fail.

“Selection then takes place, and you only select the wild type (the trait that prevails in normal conditions) [of the gene]; however, the others are still there,” says Gottlieb. “The idea is that, you are getting a panoply of variants, and they are there to protect the tissue. If you get a change in environment, they can respond to it.”

bp.blogspot.com
bp.blogspot.com

Based on this new theory, our bodies should be thought of as a composite of microenvironments to which our cells must adapt. Therefore, genetic diversity would be advantageous to an organism. Having the ability to adapt would provide cells with a better chance at survival in many different situations, such as if a virus killed off a certain cell lineage, but left another one unaffected. In order to survive, this remaining cell lineage would require genetic variation, in order to make up for the other lineage being killed off.

This theory is gaining momentum among cancer researchers, as it helps to explain our failure to identify carcinogens. For instance, cigarette smoke is believed to be a cause of lung cancer. However, when you expose cells to smoke, they are not mutagenized, compared to what occurs when you are, for example, exposed to UV rays. An alternative explanation may be that it is this smoke that is selecting for a lung cancer causing mutation, since those cancer cells thrive in the conditions. Based on this new theory, cigarette smoke does not cause a mutation; rather, the mutation was always there. The smoke selects the mutant out of our diverse gene bank.

While it is necessary that more research be conducted before drawing conclusions, it appears that the theory of natural selection should be applied to our cells. At the end of the day, why should they be treated any differently?

eyeofsky.com
a, Science & Technology

Touchscreens

Touchscreens have revolutionized the way we interact with digital devices. The most important attribute they have brought to the user experience is the reduction in the learning curve of operating a device. A simple tap on the screen can trigger commands that would have otherwise been complicated with a mouse and keyboard.

Keyboards posed a problem for some computer users, due to the fact that most populations are not 100 per cent literate. With the advent of the touchscreen, people need only touch an icon on a screen to perform the desired task, rather than type in a command. The increasing ease in operation has immense implications for increasing the universal use of technology around the world.

The idea of the touchscreen has been around for decades. The first working prototype, a resistive touchscreen, was developed by American inventor Sam Hurst in 1982. What we see in smartphones, tablets, and other devices today, however, are capacitive touchscreens. There is a marked difference in the fluidity of the touch experience between the two formats.

The resistive touchscreen is composed of transparent, electrically resistive layers, which have a thin gap between them. When the screen is pressed at a particular point, the layers come in contact, and they behave as voltage dividers. Using this mechanism, the position of the touch on the screen can be tracked and fed into the control unit.

The problems with this type of display are that it requires a harder touch to register a signal, and the mechanism wears out over time—both of which reduce the functionality of the touchscreen. In contrast, the capacitive touchscreen works on the principle that the human body is a great conductor; the screen has a glass panel coated with a transparent conductor such as indium tin oxide. When the screen is touched with the finger, the electrostatic field (a field of charged particles) of the screen is disrupted, which enables the desired function.

This process results in a highly effective phone, as the screen can be activated with merely a gentle touch. The only flaw associated with this format, is that it cannot be operated using gloves or any another material that cannot conduct (which becomes a problem during Montreal’s cold winters).

The potential applications of touchscreens are impressive. At the keynote of the 2013 Consumer Electronics Show (CES) under the theme “Mobilizing Possibilities,” Samsung introduced an innovative range of flexible displays called “Youm.” Using Organic Light Emitting Diodes (OLEDs) on a plastic sheet, they created high-resolution displays that bend. This technology allows many novel phone features, such as the expansion of the smartphone’s screen, enabling it to curve around the edges of the phone. This curved area could be used for looking at notifications, such as text messages, while the device is lying flat on a table. The Youm platform is one of many innovations that will define the next generation of touch-based devices.

a, Editorial

Editorial 101: The process behind an editorial

Every Friday, the Tribune’s editorial board meets to plan our editorial for the coming issue. We start with a range of ideas, and ultimately focus on the one that seems to us to be the most relevant, controversial, and interesting. We then discuss, each member bringing forward individual perspectives, but all the same aiming to reach group consensus. For this special issue, we take you behind the scenes of our editorial board discussion with an annotated editorial. Please see the image on the left to see the full version.

Click to see the full 'deconstructed' feature!
Click to see the full ‘deconstructed’ feature!

By the time that this editorial goes to press, the Parti Québécois’ (PQ) long-awaited Summit on Higher Education—being held on Feb. 25 and 26—will have come and gone. The Summit, announced upon the PQ’s electoral victory in September, promised open and sincere discussion on education policy that had the potential to produce immensely constructive results. Unfortunately, much of the optimism that met the initial announcement has died down; the poor communication, planning, and a range of other issues leading up to the event have greatly hampered the likelihood that the Summit will fulfill its full potential.

Although some election promises, such as the PQ’s inflexible approach to issues of language and culture, were immediately at the forefront of its agenda, the summit took much longer to materialize than most expected. Until very recently, it was only referred to by the government in the vaguest of terms: the date and location were not announced for months; parties attending the summit were only told what would be on the agenda immediately before the pre-summit consultations; even now, on the French-only website that has been put together for the event, information is extremely difficult to find, and does not show the time and location of the Summit on the front page. More than anything, the Summit looks as though it has been thrown together at the last minute, a trait that is seeming increasingly common for the PQ’s style of governance.

The nature of a minority government is such that its future is always uncertain; there is a constant need to make concessions, and satisfy enough of the opposition to remain in power. This government’s actions since being elected, however, go further than this. It seems to act with no eye to the future, simply reacting to the problem at hand. In December, the PQ announced of $140 million in funding cuts to universities across the province. They were unveiled with no prior warning to stakeholders, and shortly after the announcement of the PQ’s budget. This budget seemed to make a point not to take money from post-secondary education, and continued to herald the PQ’s implementation of a tution freeze. Once again, a lack of communication made the situation much worse than it might otherwise have been, as stakeholders scrambled to react.

Although the amount and timing of the cuts are proving catastrophic to schools—especially those like McGill that were already encountering budgetary problems—it is not all that much money for a government in the grand scheme of things. If the government viewed education as a priority, there are other places this money could have come from which would have had a less immediate and crippling impact. One example that springs to mind is the Plan Nord, which has seen none of its $2.1 billion in government funding face cuts. Ultimately, $140 million is small when dealing with billion-dollar budgets.

The cuts will inevitably shape the discussion at this week’s summit. They essentially conveyed the message, before discussions even began, that extended funding to schools is off the table. The same has been said about the possibility of a discussion on free tuition, with Minister of Higher Education Pierre Duchesne dismissing the idea only weeks before the Summit. If matters like these are not even part of the discussion, then the discussion is inevitably incomplete. The balance of power was established before the Summit even began, and the government has made it clear that it will only hear ideas to which it is already favourable.

All of this has culminated in a situation where no party approached the Summit with any real optimism as to what can be accomplished here. Whether the PQ truly doesn’t view education as a priority, or is simply finding itself overwhelmed with the realities of running a province, it has essentially doomed this Summit to failure, making little progress on deciding the future of Quebec’s education system. Although the PQ has recently seemed to equate ‘culture’ with ‘language,’ education has long been an integral part of Quebec culture. If this government chooses to neglect it, it will be writing off a large part of this province’s past—and of its future.

Carol Huynh celebrates a gold medal. (vancouverobserver.com)
a, Behind the Bench, Sports

IOC wrong to pin wrestling

As a fan of Mixed Martial Arts, it should be no surprise that I have developed an interest in the sport’s different components.

This is why I was so excited to watch Freestyle and Greco-Roman wrestling during the London 2012 Olympic games. For anyone who tuned in to watch these sports, wrestling provided some of the Olympics’ greatest moments. On the women’s side of the bracket, two Canadians, Carol Huynh and Tonya Verbeek showcased their toughness with a hard-fought bronze and silver medal, respectively. On the men’s side, the Canadian Matthew Gentry came up just short of winning a medal in the men’s 74 kg freestyle. The man to whom he lost—the sublime American wrestler, Jordan Burroughs—dominated the gold medal match in a weight class typically won by Eastern European and Iranian wrestlers.

Given the athletic drama that wrestling provides, it came as a shock that the International Olympic Committee (IOC) decided to remove wrestling from the 2020 Games last week, leaving in other sports such as modern pentathlon. Not only are many people ignorant to what modern pentathlon is, one would also be hard-pressed to find anyone who knows which five disciplines constitute this bizarre Olympic sport. For those interested, it is a combination of pistol shooting, fencing, 200m freestyle swimming, show jumping, and a three-kilometre run. Designed to emulate the plight of 19th century cavalry soldier deep in the enemy’s turf, it unquestionably captures sport in the 21st century… or not.

The IOC’s purported reasons for removing wrestling were vague and not forthcoming. They said that this decision was not intended to discredit wrestling, but rather, to affirm how important the other 25 Olympic sports are, and how the IOC now has room to include a new sport in the 2020 games.

This reasoning is foolish. First, wrestling has been included in every modern Olympics since 1896, and was the key component of the ancient Olympics. Wrestling is also a truly global sport, with 71 nations competing in the last Olympic games, and 29 different countries medalling. To eliminate this sport from the games would be a sign that the modern Olympics have no appreciation for its deep and rich tradition.

Next, the IOC’s decision reeked of corruption. Modern pentathlon maintained its status as an Olympic sport because of the persistent lobbying of the former IOC President’s son, Juan Antonio Samaranch Jr. It’s not a stretch to say that nobody would be surprised if this ‘lobbying’ took place inside of fancy hotels and with the exchanges of fancy leather briefcases.

Wrestling also does not have a professional league. The Olympics is the only way any wrestler can ever expect to make a living from the sport, an option the IOC’s decision has put in jeopardy.

Finally, the move is unnecessary as the IOC could have just extended the Olympics another day, without eliminating any sport, and still adding new ones.

Fortunately, all is not lost for aficionados of this noble sport. Wrestling still has a chance of being picked for the 2020 games amongst a host of other sports such asbvnvgbnn   wakeboarding, karate, golf, baseball, squash, wushu, and sports climbing. Countries like Iran, Russia, and the United States are all deeply invested in maintaining wrestling’s status as an Olympic sport, and will likely lobby very hard on its behalf. If, for some reason, the IOC does not reinstate wrestling into the games, the committee will be sending a very clear message: consumerism and special interests are more important than the pursuit of athletic excellence.

 

Read the latest issue

Read the latest issue