Questions about subtitles and language learning

Posted: August 28, 2020 in research, video, vocabulary
Tags: , , , , , , , , ,

subtitlesAs both a language learner and a teacher, I have a number of questions about the value of watching subtitled videos for language learning. My interest is in watching extended videos, rather than short clips for classroom use, so I am concerned with incidental, rather than intentional, learning, mostly of vocabulary. My questions include:

  • Is it better to watch a video that is subtitled or unsubtitled?
  • Is it better to watch a video with L1 or L2 subtitles?
  • If a video is watched more than once, what is the best way to start and proceed? In which order (no subtitles, L1 subtitles and L2 subtitles) is it best to watch?

For help, I turned to three recent books about video and language learning: Ben Goldstein and Paul Driver’s Language Learning with Digital Video (CUP, 2015), Kieran Donaghy’s Film in Action (Delta, 2015) and Jamie Keddie’s Bringing Online Video into the Classroom (OUP, 2014). I was surprised to find no advice, but, as I explored more, I discovered that there may be a good reason for these authors’ silence.

There is now a huge literature out there on subtitles and language learning, and I cannot claim to have read it all. But I think I have read enough to understand that I am not going to find clear-cut answers to my questions.

The learning value of subtitles

It has been known for some time that the use of subtitles during extensive viewing of video in another language can help in the acquisition of that language. The main gains are in vocabulary acquisition and the development of listening skills (Montero Perez et al., 2013). This is true of both L1 subtitles (with an L2 audio track), sometimes called interlingual subtitles, (Incalcaterra McLoughlin et al, 2011) and L2 subtitles (with an L2 audio track), sometimes called intralingual subtitles or captions (Vanderplank, 1988). Somewhat more surprisingly, vocabulary gains may also come from what are called reversed subtitles (L2 subtitles and an L1 audio track) (Burczyńska, 2015). Of course, certain conditions apply for subtitled video to be beneficial, and I’ll come on to these. But there is general research agreement (an exception is Karakaş & Sariçoban, 2012) that more learning is likely to take place from watching a subtitled video in a target language than an unsubtitled one.

Opposition to the use of subtitles as a tool for language learning has mostly come from three angles. The first of these, which concerns L1 subtitles, is an antipathy to any use at all of L1. Although such an attitude remains entrenched in some quarters, there is no evidence to support it (Hall & Cook, 2012; Kerr, 2016). Researchers and, increasingly, teachers have moved on.

The second reservation that is sometimes expressed is that learners may not attend to either the audio track or the subtitles if they do not need to. They may, for example, ignore the subtitles in the case of reversed subtitles or ignore the L2 audio track when there are L1 subtitles. This can, of course, happen, but it seems that, on the whole, this is not the case. In an eye-tracking study by Bisson et al (2012), for example, it was found that most people followed the subtitles, irrespective of what kind they were. Unsurprisingly, they followed the subtitles more closely when the audio track was in a language that was less familiar. When conditions are right (see below), reading subtitles becomes a very efficient and partly automatized cognitive activity, which does not prevent people from processing the audio track at the same time (d’Ydewalle & Pavakanun, 1997).

Related to the second reservation is the concern that the two sources of information (audio and subtitles), combined with other information (images and music or sound effects), may be in competition and lead to cognitive overload, impacting negatively on both comprehension and learning. Recent research suggests that this concern is ungrounded (Kruger et al, 2014). L1 subtitles generate less cognitive load than L2 subtitles, but overload is not normally reached and mental resources are still available for learning (Baranowska, 2020). The absence of subtitles generates more cognitive load.

Conditions for learning

Before looking at the differences between L1 and L2 subtitles, it’s a good idea to look at the conditions under which learning is more likely to take place with subtitles. Some of these are obvious, others less so.

First of all, the video material must be of sufficient intrinsic interest to the learner. Secondly, the subtitles must be of a sufficiently high quality. This is not always the case with automatically generated captions, especially if the speech-to-text software struggles with the audio accent. It is also not always the case with professionally produced L1 subtitles, especially when the ‘translations are non-literal and made at the phrase level, making it hard to find connections between the subtitle text and the words in the video’ (Kovacs, 2013, cited by Zabalbeascoa et al., 2015: 112). As a minimum, standard subtitling guidelines, such as those produced for the British Channel 4, should be followed. These limit, for example, the number of characters per line to about 40 and a maximum of two lines.

For reasons that I’ll come on to, learners should be able to switch easily between L1 and L2 subtitles. They are also likely to benefit if reliably accurate glosses or hyperlinks are ‘embedded in the subtitles, making it possible for a learner to simply click for additional verbal, auditory or even pictorial glosses’ (Danan, 2015: 49).

At least as important as considerations of the materials or tools, is a consideration of what the learner brings to the activity (Frumuselu, 2019: 104). Vanderplank (2015) describes these different kinds of considerations as the ‘effects of’ subtitles on a learner and the ‘effects with’ subtitles on learner behaviour.

In order to learn from subtitles, you need to be able to read fast enough to process them. Anyone with a slow reading speed (e.g. some dyslexics) in their own language is going to struggle. Even with L1 subtitles, Vanderplank (2015: 24) estimates that it is only around the age of 10 that children can do this with confidence. Familarity with both the subject matter and with subtitle use will impact on this ability to read subtitles fast enough.

With L2 subtitles, the language proficiency of the learner related to the level of difficulty (especially lexical difficulty) of the subtitles will clearly be of some significance. It is unlikely that L2 subtitles will be of much benefit to beginners (Taylor, 2005). It also suggests that, at lower levels, materials need to be chosen carefully. On the whole, researchers have found that higher proficiency levels correlate with greater learning gains (Pujadas & Muñoz, 2019; Suárez & Gesa, 2019), but one earlier meta-analysis (Montero Perez et al., 2013) did not find that proficiency levels were significant.

Measures of general language proficiency may be too blunt an instrument to help us all of the time. I can learn more from Portuguese than from Arabic subtitles, even though I am a beginner in both languages. The degree of proximity between two languages, especially the script (Winke et al., 2010), is also likely to be significant.

But a wide range of other individual learner differences will also impact on the learning from subtitles. It is known that learners approach subtitles in varied and idiosyncratic ways (Pujolá, 2002), with some using L2 subtitles only as a ‘back-up’ and others relying on them more. Vanderplank (2019) grouped learners into three broad categories: minimal users who were focused throughout on enjoying films as they would in their L1, evolving users who showed marked changes in their viewing behaviour over time, and maximal users who tended to be experienced at using films to enhance their language learning.

Categories like these are only the tip of the iceberg. Sensory preferences, personality types, types of motivation, the impact of subtitles on anxiety levels and metacognitive strategy awareness are all likely to be important. For the last of these, Danan (2015: 47) asks whether learners should be taught ‘techniques to make better use of subtitles and compensate for weaknesses: techniques such as a quick reading of subtitles before listening, confirmation of word recognition or meaning after listening, as well as focus on form for spelling or grammatical accuracy?’

In short, it is, in practice, virtually impossible to determine optimal conditions for learning from subtitles, because we cannot ‘take into account all the psycho-social, cultural and pedagogic parameters’ (Gambier, 2015). With that said, it’s time to take a closer look at the different potential of L1 and L2 subtitles.

L1 vs L2 subtitles

Since all other things are almost never equal, it is not possible to say that one kind of subtitles offers greater potential for learning than another. As regards gains in vocabulary acquisition and listening comprehension, there is no research consensus (Baranowska, 2020: 107). Research does, however, offer us a number of pointers.

Extensive viewing of subtitled video (both L1 and L2) can offer ‘massive quantities of authentic and comprehensible input’ (Vanderplank, 1988: 273). With lower level learners, the input is likely to be more comprehensible with L1 subtitles, and, therefore, more enjoyable and motivating. This makes them often more suitable for what Caimi (2015: 11) calls ‘leisure viewing’. Vocabulary acquisition may be better served with L2 subtitles, because they can help viewers to recognize the words that are being spoken, increase their interaction with the target language, provide further language context, and increase the redundancy of information, thereby enhancing the possibility of this input being stored in long-term memory (Frumuselu et al., 2015). These effects are much more likely with Vanderplank’s (2019) motivated, ‘maximal’ users than with ‘minimal’ users.

There is one further area where L2 subtitles may have the edge over L1. One of the values of extended listening in a target language is the improvement in phonetic retuning (see, for example, Reinisch & Holt, 2013), the ability to adjust the phonetic boundaries in your own language to the boundaries that exist in the target language. Learning how to interpret unusual speech-sounds, learning how to deal with unusual mappings between sounds and words and learning how to deal with the acoustic variations of different speakers of the target language are all important parts of acquiring another language. Research by Mitterer and McQueen (2009) suggests that L2 subtitles help in this process, but L1 subtitles hinder it.

Classroom implications?

The literature on subtitles and language learning echoes with the refrain of ‘more research needed’, but I’m not sure that further research will lead to less ambiguous, practical conclusions. One of my initial questions concerned the optimal order of use of different kinds of subtitles. In most extensive viewing contexts, learners are unlikely to watch something more than twice. If they do (watching a recorded academic lecture, for example), they are likely to be more motivated by a desire to learn from the content than to learn language from the content. L1 subtitles will probably be preferred, and will have the added bonus of facilitating note-taking in the L1. For learners who are more motivated to learn the target language (Vanderplank’s ‘maximal’ users), a sequence of subtitle use, starting with the least cognitively challenging and moving to greater challenge, probably makes sense. Danan (2015: 46) suggests starting with an L1 soundtrack and reversed (L2) subtitles, then moving on to an L2 soundtrack and L2 subtitles, and ending with an L2 soundtrack and no subtitles. I would replace her first stage with an L2 soundtrack and L1 subtitles, but this is based on hunch rather than research.

This sequencing of subtitle use is common practice in language classrooms, but, here, (1) the video clips are usually short, and (2) the aim is often not incidental learning of vocabulary. Typically, the video clip has been selected as a tool for deliberate teaching of language items, so different conditions apply. At least one study has confirmed the value of the common teaching practice of pre-teaching target vocabulary items before viewing (Pujadas & Muñoz, 2019). The drawback is that, by getting learners to focus on particular items, less incidental learning of other language features is likely to take place. Perhaps this doesn’t matter too much. In a short clip of a few minutes, the opportunities for incidental learning are limited, anyway. With short clips and a deliberate learning aim, it seems reasonable to use L2 subtitles for a first viewing, and no subtitles thereafter.

An alternative frequent use of short video clips in classrooms is to use them as a springboard for speaking. In these cases, Baranowska (2020: 113) suggests that teachers may opt for L1 subtitles first, and follow up with L2 subtitles. Of course, with personal viewing devices or in online classes, teachers may want to exploit the possibilities of differentiating the subtitle condition for different learners.

REFERENCES

Baranowska, K. (2020). Learning most with least effort: subtitles and cognitive load. ELT Journal 74 (2): pp.105 – 115

Bisson, M.-J., Van Heuven, W.J.B., Conklin, K. and Tunney, R.J. (2012). Processing of native and foreign language subtitles in films: An eye tracking study. Applied Psycholingistics, 35 (2): pp. 399 – 418

Burczyńska, P. (2015). Reversed Subtitles as a Powerful Didactic Tool in SLA. In Gambier, Y., Caimi, A. & Mariotti, C. (Eds.), Subtitles and Language Learning. Principles, strategies and practical experiences. Bern: Peter Lang (pp. 221 – 244)

Caimi, A. (2015). Introduction. In Gambier, Y., Caimi, A. & Mariotti, C. (Eds.), Subtitles and Language Learning. Principles, strategies and practical experiences. Bern: Peter Lang (pp. 9 – 18)

Danan, M. (2015). Subtitling as a Language Learning Tool: Past Findings, Current Applications, and Future Paths. In Gambier, Y., Caimi, A. & Mariotti, C. (Eds.), Subtitles and Language Learning. Principles, strategies and practical experiences. Bern: Peter Lang (pp. 41 – 61)

d’Ydewalle, G. & Pavakanun, U. (1997). Could Enjoying a Movie Lead to Language Acquisition?. In: Winterhoff-Spurk, P., van der Voort, T.H.A. (Eds.) New Horizons in Media Psychology. VS Verlag für Sozialwissenschaften, Wiesbaden. https://doi.org/10.1007/978-3-663-10899-3_10

Frumuselu, A.D., de Maeyer, S., Donche, V. & Gutierrez Colon Plana, M. (2015). Television series inside the EFL classroom: bridging the gap between teaching and learning informal language through subtitles. Linguistics and Education, 32: pp. 107 – 17

Frumuselu, A. D. (2019). ‘A Friend in Need is a Film Indeed’: Teaching Colloquial Expressions with Subtitled Television Series. In Herrero, C. & Vanderschelden, I. (Eds.) Using Film and Media in the Language Classroom. Bristol: Multimedia Matters. pp.92 – 107

Gambier, Y. (2015). Subtitles and Language Learning (SLL): Theoretical background. In Gambier, Y., Caimi, A. & Mariotti, C. (Eds.), Subtitles and Language Learning. Principles, strategies and practical experiences. Bern: Peter Lang (pp. 63 – 82)

Hall, G. & Cook, G. (2012). Own-language Use in Language Teaching and Learning. Language Learning, 45 (3): pp. 271 – 308

Incalcaterra McLoughlin, L., Biscio, M. & Ní Mhainnín, M. A. (Eds.) (2011). Audiovisual Translation, Subtitles and Subtitling. Theory and Practice. Bern: Peter Lang

Karakaş, A. & Sariçoban, A. (2012). The impact of watching subtitled animated cartoons on incidental vocabulary learning of ELT students. Teaching English with Technology, 12 (4): pp. 3 – 15

Kerr, P. (2016). Questioning ‘English-only’ Classrooms: Own-language Use in ELT. In Hall, G. (Ed.) The Routledge Handbook of English Language Teaching (pp. 513 – 526)

Kruger, J. L., Hefer, E. & Matthew, G. (2014). Attention distribution and cognitive load in a subtitled academic lecture: L1 vs. L2. Journal of Eye Movement Research, 7: pp. 1 – 15

Mitterer, H. & McQueen, J. M. (2009). Foreign Subtitles Help but Native-Language Subtitles Harm Foreign Speech Perception. PLoS ONE 4 (11): e7785.doi:10.1371/journal.pone.0007785

Montero Perez, M., Van Den Noortgate, W., & Desmet, P. (2013). Captioned video for L2 listening and vocabulary learning: A meta-analysis. System, 41, pp. 720–739 doi:10.1016/j.system.2013.07.013

Pujadas, G. & Muñoz, C. (2019). Extensive viewing of captioned and subtitled TV series: a study of L2 vocabulary learning by adolescents, The Language Learning Journal, 47:4, 479-496, DOI: 10.1080/09571736.2019.1616806

Pujolá, J.- T. (2002). CALLing for help: Researching language learning strategies using help facilities in a web-based multimedia program. ReCALL, 14 (2): pp. 235 – 262

Reinisch, E. & Holt, L. L. (2013). Lexically Guided Phonetic Retuning of Foreign-Accented Speech and Its Generalization. Journal of Experimental Psychology: Human Perception and Performance. Advance online publication. doi: 10.1037/a0034409

Suárez, M. & Gesa, F. (2019) Learning vocabulary with the support of sustained exposure to captioned video: do proficiency and aptitude make a difference? The Language Learning Journal, 47:4, 497-517, DOI: 10.1080/09571736.2019.1617768

Taylor, G. (2005). Perceived processing strategies of students watching captioned video. Foreign Language Annals, 38(3), pp. 422-427

Vanderplank, R. (1988). The value of teletext subtitles in language learning. ELT Journal, 42 (4): pp. 272 – 281

Vanderplank, R. (2015). Thirty Years of Research into Captions / Same Language Subtitles and Second / Foreign Language Learning: Distinguishing between ‘Effects of’ Subtitles and ‘Effects with’ Subtitles for Future Research. In Gambier, Y., Caimi, A. & Mariotti, C. (Eds.), Subtitles and Language Learning. Principles, strategies and practical experiences. Bern: Peter Lang (pp. 19 – 40)

Vanderplank, R. (2019). ‘Gist watching can only take you so far’: attitudes, strategies and changes in behaviour in watching films with captions, The Language Learning Journal, 47:4, 407-423, DOI: 10.1080/09571736.2019.1610033

Winke, P., Gass, S. M., & Sydorenko, T. (2010). The Effects of Captioning Videos Used for Foreign Language Listening Activities. Language Learning & Technology, 1 (1): pp. 66 – 87

Zabalbeascoa, P., González-Casillas, S. & Pascual-Herce, R. (2015). In Gambier, Y., Caimi, A. & Mariotti, C. (Eds.), Subtitles and Language Learning. Principles, strategies and practical experiences Bern: Peter Lang (pp. 105–126)

Comments
  1. Marc says:

    Hi Philip,

    Really good, comprehensive post as usual. I think this can be really complicated as well because there is an indication that shared orthography can hinder phonological acquisition, especially phoneme length (Hayes-Harb et al. 2010) and that in some cases it can help phonological acquisition of lexis (Showalter 2016, I believe). I suppose the quick answer is that context is important, as is first/second/other languages.

    I do know that my students sometimes overuse subtitles as a crutch, but – like you say, more constructive use of the subtitles needs to be taught.

    At a conference last year, Natalia Wisniewska presented work on pronunciation acquisition for L2 English among (primarily?) L1 Catalan and Spanish learners based on L1/L2 subtitles. L1 subtitles help meaning acquisition, L2 subtitles pronunciation. There was no effect found for listening. There’s a paper up now: https://www.cambridge.org/core/journals/studies-in-second-language-acquisition/article/can-captioned-video-benefit-second-language-pronunciation/77D4AA3A5C6EBF6800157C1BAA00149F . Not open access but there are ways and means 😉

    • philipjkerr says:

      Thanks, Marc. I’ll have to find some ways and means to read that. What intrigues me (in the paper you link to) is the difference in findings (phonological accuracy in perception) between this research and other. It really all is rather complicated, isn’t it?

      • Marc says:

        It certainly is, and I guess that is one of the reasons that it is hard for teachers to engage with it – there are several seeming contradictions but it’s really a complex state of considering priorities for all 30+ learners in a class, that may or may not be like the participants in the research study. Anyway, I could chin stroke and ponder all night, so will stop here!

  2. eflnotes says:

    Thanks Philip great post, I think you rightly highlight the importance of individual differences which implies learner control over how to use subtitles.

    Did you come across any research looking at using full subtitles and partial subtitles? I ask because I quite like the idea of partial synchronized captions, where selected words are shown. You can see example here https://interact.englishup.me/PSC-example/.

    The authors are working on a learner adapted version – the system uses words learners want to show or hide.
    ta
    mura

    • philipjkerr says:

      Hi Mura,
      That’s interesting, thanks. I hadn’t come across this before … I missed the developers’ article in reCALL a few years ago.
      It’s a little like speed reading, but a bit weird, I think. The example in the link you give is perhaps not the best example that could be chosen of the technique at work. The speaker, Dan Pink, delivers the talk at a fair old rattle.
      The captions that are shown seem to be words with heavy stress, so mostly lexical content words, some low frequency, but some high. As you watch, it’s hard not to try to fill in the gaps, but it’s hard to do so. And I, for one, felt that I was getting cognitively overloaded, so this partial caption technique may be more useful as a viewing option after I’ve already watched once with full captions.
      It would be interesting to see how a learner adaptable version would work. How would a learner know which words they want to show or hide?
      I’ve now had a look at the article (Mirzaei, M.S., Kourosh, M., Akita, Y. & Kawahara, T. (2017). Partial and synchronized captioning: A new tool to assist learners in developing second language listening skill. ReCALL, 29 (2): 178 – 199). The authors claim that their approach leads to the same level of comprehension as full captions, and can help to wean learners off full captions. I’d like to see more replication, especially with different kinds of students (those here were Japanese undergraduates of different levels of proficiency, but typically with ten years of English study behind them). Still, I’m more than happy to accept the results for the time being.
      The authors are interested in facilitating access to comprehensible input, but they don’t show any interest in the vocabulary gains that seem to come from full-caption use. It’s hard to see how these partial captions could be very useful for lexical development … but I may be wrong.
      Thanks again – it was interesting to learn more about this.
      Philip

      • eflnotes says:

        hi again re learner adapted one here is their paper on that – Learner-adaptive partial and synchronized captions for L2 listening skill development [https://research-publishing.net/manuscript?10.14705/rpnet.2019.38.1025]

        a 2019 paper talks about a learner’s “caption reliance” and how this can moderate effects of full, partial and real-time captions (real time captions seems closer to the speed reading effect you mention) – One size fits all? In search of the desirable caption display for second language learners with different caption reliance in listening comprehension [https://journals.sagepub.com/doi/10.1177/1362168819856451]

  3. philipjkerr says:

    Thanks again, Mura.
    I had a good look at the first paper you cite (Mirzaei, M. S., & Meshgi, K. (2019). Learner-adaptive partial and synchronized captions for L2 listening skill development. In F. Meunier, J. Van de Vyver, L. Bradley & S. Thouësny (Eds), CALL and complexity – short papers from EUROCALL 2019 (pp. 291-296)). It appears that, with the individualised captions, users are given a video with automated partial captions and they are asked to click on spaces where they wanted to see the word, and to click on words which they didn’t need to see. This is then used as training data for the system to automatically select which items will be shown / not shown in future partially captioned videos. I have to say that this leaves me skeptical. The conditions under which the training data is generated are very odd (just think of the cognitive load that is required!). Of course, as the authors discuss, the process of generating training data can be repeated, resulting in better auto-selection of revealed / non-revealed items. But this entails a pretty big commitment / high motivation on the part of the learner, which (1) cannot be assumed in many (most?) contexts, and (2) may not be worth the effort anyway.
    I say that it may not be worth the effort anyway, because the authors say that their main reason for developing the system is because ‘many learners, especially beginners, struggle with cognitive load and split attention, while attending to caption text together with other modes of input’. But beginners probably shouldn’t be using L2 captions in the first place; they’d be better off with L1 subtitles. And, for others, other research (e.g. Baranowska) suggests that cognitive overload isn’t a problem. I can see that partial captions may be useful for some learners on some occasions, but I can’t help feeling that automated partial captions may be a ‘solution’ to a non-problem.
    The other article (Lee, P. J., Liu, Y. T., & Tseng, W. T. (Accepted/In press). One size fits all? In search of the desirable caption display for second language learners with different caption reliance in listening comprehension. Language Teaching Research. https://doi.org/10.1177/1362168819856451) looks interesting, too. I can’t find open access to the full article, but the main claim seems to be that learners who are less reliant on captions benefit more (than more caption-reliant learners) from partial captions and less from full captions. I assume that caption-reliance must correlate quite strongly with proficiency, so the first part of this finding is not surprising. More caption-reliant learners benefit more from full captions. The conclusion would, then, be that partial captions are of most use to learners who need them least. And this is not exactly a ringing endorsement of partial-caption use.
    Anyway, interesting stuff, but the take-away is that one size won’t fit all.
    Thanks again
    Philip

  4. eflnotes says:

    valid points Philip thx;

    re learner adaptive captions – i think may not be so relevant in the context of incidental learning that you have focused on in this blog, as you imply if you want to watch something without worrying about learning any language then the adaptive system would get in the way;

    the second study (available thru sci-hub) controlled for proficiency – learners were B2 overall and also B2 on listening skills

  5. Sandy Millin says:

    Another fascinating post Philip – thank you! And thanks to Marc and Mura for the comment contributions too.

    “Somewhat more surprisingly, vocabulary gains may also come from what are called reversed subtitles (L2 subtitles and an L1 audio track) (Burczyńska, 2015).”
    I’m not surprised by this at all – I regularly go to the cinema in foreign countries and learn one or two words from the L2 subtitles almost every time I watch a film in English (or French or any other language I know). Seeing that word or phrase again later, I often remember the film I was watching when I learnt it. Sometimes the L2 subtitles can be handy too when the speakers have a very strong accent or are mumbling too much in English – there have been times (even when I was pre-int) when the L2 subtitles were the only reason I knew what was going on! That’s all anecdotal, and I’m a very experienced language learner, a fast reader and a passionate film fan, so no idea if I’m the norm or the exception here.

    Sandy

  6. frdterry says:

    I do content for learners/gamers who would like to be able to play popular video games with native English speakers. I find that short clips featuring a wide variety of games and accents work well, starting without (English) subtitles, then adding subtitles and ending with one run-through without. A big advantage with game clips is that the learners can see what is happening. It seems to me for longer extracts the approach needs to be different, focusing more on preparation for listening to an extended text – with subtitles of course. I’m currently working on a seven-minute video with non-stop unscripted commentary from a gamer doing a 10-minute house-building challenge in Sims 4. It is sort of a top of the mountain activity — it doesn’t get more challenging than this — and I think is really helpful to show learners what to expect.

    Terry Fredrickson
    Real English for Gamers
    https://www.youtube.com/channel/UCvRkPfGnOm_lxAr7SyWBJ8g

Leave a Reply

Fill in your details below or click an icon to log in:

WordPress.com Logo

You are commenting using your WordPress.com account. Log Out /  Change )

Google photo

You are commenting using your Google account. Log Out /  Change )

Twitter picture

You are commenting using your Twitter account. Log Out /  Change )

Facebook photo

You are commenting using your Facebook account. Log Out /  Change )

Connecting to %s