The TR, Exegesis, & Hermeneutics

Introduction

The most difficult bridge to gap with participants in this discussion is convincing them that this is more than just textual science. I have always emphasized this on my blog, that this is topic is deeply theological. In the tradition of my spiritual forefathers, all loci of study must not only be theological, but also practical. We do not needlessly debate the quantity of angels that can dance on the pin of a needle. One of the greatest errors of many Modern Critical Text advocates is to partition this part of the discussion into science and faith. Often, the discussion is made to seem like the textual scholars are doing hard science which is then handed over to Christians for interpretation and application. This is why many advocates of the modern text do not understand when those in the TR camp begin the debate with theological foundations rather than text critical ones. As a result of the “scientification” of the text in its creation, the same sterile techniques seem to have been transferred into the handling of the Word.

Mishandling of the Text Through Atomization

Perhaps it is the case that these things are not connected, but it is true that modern methods of textual scholarship have arrived hand in hand with modern exegesis and hermeneutics in the seminaries. This is why, in my opinion, modern reformed and reformed adjacent preaching is dry and difficult to understand. It is also often alienating to the layman. The method often taught at seminaries today is called the “historical grammatical” approach. An internet search will tell you this is a “Protestant” method of exegesis, though I would argue that there is nothing historically protestant about it. This method attempts to steer the exegete towards the true meaning of the text by focusing on grammar, syntax, and historical context. At face value, this sounds great, but in it’s application it often falls short. Such is the case with many well-meaning efforts to adapt scholarly methods for the church. I will say, it does do a wonderful job in steering pastors away from sports references or sermons filled to the brim with anecdotes and analogies.

First, I want to point out that this method has a historical critical trojan horse built into it at the outset. While the faithful minister would never think of doing this, a man pulled by the academy might consider “historical context” a license to inject critical perspectives into the text. Historical context, for example, might be a new perspective on Paul, or perhaps applying modern gender theory to the pericope adulterae as we see in Jennifer Knust and Elijah Hixson’s book “To Cast the First Stone.” Even if the exegete does not consider such extreme examples, the method itself opens the door to changing historical perspectives on the text. I will admit, this exegetical error is far more common in the less conservative corners of the Reformed world, but it is present in more benign applications.

What is far more common in conservative circles are exegetical errors produced in the grammatical portion of this method. While it is true that understanding the text properly requires accurate analysis of the grammar and syntax of a text, the way that this is applied is oftentimes harmful to the text. The primary error that many exegetes make is to handle the words of the text outside of the text itself. I remember vaguely in my early Christian years being amazed by a pastor who preached an entire sermon on the Greek word, “The.” I look back on that now, thinking how absurd that is. This of course is the most extreme possible example, but many such cases exist as it pertains to words like love, finished, and so on. This happens when the exegete doesn’t treat Greek and Hebrew like a real language, but a mystical one. I cannot count the number of times I’ve heard a pastor in a Reformed church say something along the lines of, “In English, this word means this, but in Greek, it means THIS THIS.” In almost every case of this happening, the pastor is merely taking a synonym of the Greek word, and emphatically presenting the word to the church as though they don’t understand how synonyms work. One egregious case of this is when John MacArthur famously translated the word, “run” to “glide” and gave the impression that somebody was gliding along the ground.

This kind of exegesis occurs when the exegete fails to understand what language is in the first place, not just Greek or Hebrew. The reality is, most adult Americans read at a 3rd or 4th grade level, so this kind of error is not unsurprising. What must be corrected is the idea that there is magic in the middle of translating one language to another. It is often said that two languages can never be perfectly translated from one to the other. This is true, if we are saying that every word must be translated to an equivalent word in the target language. I recently got into a heated debate with one of my friends who knows Japanese and English. He told me, “It is impossible to translate Japanese into English perfectly.” So I said, “I bet you’re wrong, give me an example.” He brought up a word that had a complex meaning, but he was able to perfectly articulate it into English for me such that I now understood what the word meant. After he described what the word meant, I responded, “See, it can be translated.” What he actually meant was that the word did not map to a single English word, but it definitely could be translated. This is why “word for word” translations use a flawed translation methodology because translation is not simply “word for word.”

This type of realization is obvious to anybody who has actually learned a second language, especially if the language uses conjugations and declensions to change the word into different forms, like Greek. Much of these types of errors occur because seminary students do not study Greek as a language. Seminary students learn just enough vocabulary to use language tools, but could not go to Greece and order a meal. Daniel Street proved this when he tested seminary Greek professors on 10 basic vocabulary words, and they all failed. Seminaries are not teaching Greek as a language, and it is damaging the pulpit. The reality is, the Bible is not written in “ancient Greek,” it is written in a form of Greek that is about as far from modern Greek as King James English is from us today. You can take a Greek New Testament, open it up in Greece, read it out loud, and every single elementary school child will be able to understand what it is saying. They may think you sound funny, but they will understand you.

The point in saying all of that is to say this, the grammatical portion of modern exegetical methods sounds great, but it falls apart if the exegete does not understand Greek as a language, or even language learning in general. I will say this, if the exegete cannot go to Greece and have a conversation with the people there, he is not learned enough to utilize the Greek of the New Testament without falling into unfortunate errors. Believe it or not, you must know a language to engage with that language in literary form. Let me provide an analogy to drive my point home. Imagine you take a Spanish speaker, teach him English for 1-2 years, and then tell him to do advanced literary analysis of Shakespeare in English. That is what our Seminaries do for Greek language studies.

The end result of this method is that pastors, who have less understanding of Greek than a 2nd grader in Greece, are doing grammatical and syntactical analysis in a language they barely understand, and coming away with meanings that are often nonsensical or exactly the same as the English. This is why when pastors appeal to the original languages in a sermon, they often mysticise the text, as if the true meaning cannot be discovered in the English translation. Even more common, is for a pastor to simply say the Greek word (often incorrectly or in a pronunciation that nobody in history would recognize), and then say the definition in English, which their translation has already done for them! They do this because they don’t understand how translation from any language works, and they also do not really know the language they are trying to translate. The effect of hyper focusing on grammar and syntax is that it forces men into language analysis in a language that cannot be learned in 4 semesters. It would be ridiculous to ask such things from somebody learning English as a second language, so why do we impose this upon those pursuing the pulpit?

Unless the exegete actually knows Greek and Hebrew, he should be analyzing the grammar and syntax of the translation he has before him in the language he can understand. I will say, in every single case I have ever heard of a pastor doing Greek or Hebrew analysis from the pulpit, the same meaning could have easily been derived by simply analyzing the text in English and using an English dictionary. Now, I do think it is incredibly important for pastors to learn Greek and Hebrew, but that effort takes more than what your seminary has to offer. What is unfortunate about all of this, is that men who do not know Greek are taught to use Greek in their sermons. The effect is that the average layperson now believes that their translation is inadequate to understand the Word of God. Every Greek word is made to seem as though it has 50 meanings, and only the pastor truly knows what it means. It is, ironically, a modern practice of having secret knowledge, and a chief example of private interpretation. There is a correct way to use the Greek in sermon preparation, but that involves actually knowing the language.

The modern method of exegesis teaches pastors to look at a text in a language they do not know, pull words out of the sentence, look at those words individually, and place them back into the text pregnant with meaning. They then take this meaning and present to the congregation in a sermon, and the people of God are impressed that their pastor used a Greek lexicon or dictionary or language tool. More importantly, the layperson believes that in order to understand the Bible, they must wait for the man of God to tell them what it says, or go back to the Greek themselves.

What does this have to do with the TR?

One of the reasons that people take the TR position is that they see it as a logical consequence of returning to the old paths. While many people turn back to Rome, or perhaps the Eastern Orthodox church, those that wish to stay “reformed” see the TR view as a historical artifact of the Reformation. In many ways, this view is a rejection of modernity that extends beyond the textual debate. Perhaps it is as simple as people being tired with the chronological snobbery of the scholar class. In any case, there is value in doing things the way our spiritual fathers in the faith did them. In my opinion, instead of the historical grammatical approach, the seminaries should be teaching Perkins and Puritan preaching. Instead of teaching diagrams of Greek syntax, seminaries should be teaching students how to exegete the text in their mother tongue. Instead of focusing on two years of Greek vocab and grammar, seminaries should teach students conversational Greek with the eventual goal of mastery.

In any case, I am an advocate for returning to the old paths. We, as Reformed people, have treasures upon treasures in our spiritual heritage. There is no need to reinvent the wheel every decade.

Modern Scholarship, Textual Variation, and the Received Text

Dr. Peter Gurry published a short article entitled Cole on Preservation and the Westminster Divines’ Sermons back in January of this year. This is an addendum to his academic article posted in the Midwestern Journal of Theology called “Textual Criticism in the Reformation.” This is a part of the continued effort to support the claim that modern textual scholars are doing the same thing as the Reformed during the Reformation and post-Reformation era, “These arguments are the all-important context for understanding the practice of textual criticism in the Reformation” (pg. 20). He notes correctly that this is not a term the Reformers used, which is of course true because the concept of critical studies was yet to be born. He also does a great job of pulling from Milne, Muller, and Turretin to give appropriate context, which many of his colleagues avoid. He cites Muller, saying that the Reformed did not “seek the infinite regress of the lost autographa as the prop for textual infallibility” (pg. 27). He rightly notices that the Reformed considered the copies of Scripture to be authentic as the originals.

Here is where, I think, the modern textual scholars get the argument confused. The tendency seems to be to anachronistically apply the term, “Textual Criticism” to the efforts of the Reformed and post-Reformed. Where many popular level apologists go wrong is to make the claim that the giants of our tradition were ignorant of textual variation in the manuscript tradition. Gurry does not make this error as his peers often do. He says, “None of this means that scholars and theologians in this period were unaware of variants within Greek manuscripts. One can find discussion of them in Erasmus and Beza, of course, and in plenty of other scholars’ too” (pg. 29). This of course is true, and these discussions continue all the way through Turretin, Owen, Gill, and more. Dr. Gurry even invokes Jan Krans correctly in his analysis of Erasmus, noting that Erasmus was engaging in some proto-genealogical method. An important addition is to remember that this actually distinguishes Erasmus from Beza, who was far too theologically motivated for Krans’ liking.

One important observation here is that Dr. Gurry invokes one of Turretin’s questions, “Are the sources so pure that no fault has crept into the sacred manuscripts, either through the waste of time, the carelessness of copyists or the malice of the Jews or of heretics?” (Francis Turretin, Institutes of Elenctic Theology, Vol. 1, 106). He correctly characterizes the Reformed position via Turretin, which is to answer, “no.” This is where Dr. Gurry gets off the Reformed boat. He points out that modern evidence makes such claims such as the inspiration of the vowel points, “impossible to make” (pg. 33). He uses this as a starting point to argue that the Reformed could not have had a complete theology because they did not have a complete set of evidence.

This is the crux of the difference between the evangelical textual scholars and the non-evangelical textual scholars – the evangelicals appreciate the work of the Reformed and take their words into consideration. However, at the end of the day, they ultimately conclude that modern evidence has discredited the Reformed conclusions. Yet the question remains as to whether or not the Reformed would have consented to the conclusions of modernity. The modern scholars obviously say, “yes, how could they not?” Those in the Confessional Text camp say, “I don’t think they would.”

Would the Reformed have valued the discovery of the Dead Sea Scrolls as highly as modern scholars do? Would they likewise value Vaticanus and Sinaiticus as highly as modern scholars do? Dr. Gurry acknowledges that the Reformed were aware of textual variation, and yet they denied corruption, unlike modern textual scholars. The reason this is important, is because if the modern textual scholars are correct about the evidence discovered in the 19th and 20th centuries, the first chapter of the Reformed confessions must be understood in a different light and even reformulated. The evangelical textual scholars understand this, as they have produced several treatises and statements on the doctrine of Scripture in the last several years (Jongkind, Wallace, Brash). If chapter 1 of the confessions were adequate in light of modern evidence, there would be no need to add to the formulations produced in the Reformation and Post-Reformation period.

The reality is, most modern Reformed have already adopted the conclusions of modern textual scholars regarding the vowel points, the textual evidence, and the doctrinal statements regarding Scripture. The question that I always ask myself is this, “Would the Reformed have understood this evidence differently than our modern scholars?” My answer of course is, “yes.” Ultimately, this work by Dr. Gurry indicates a shift in the modern textual argument from, “The Reformed agree with us” to “The Reformed laid the foundations, but were ultimately betrayed by their lack of evidence.”

I see this as a welcome update to the conversation, which has been difficult to navigate due to modern scholars claiming the Reformed as their own for the past several years. The impact of this debate proliferates down to the layman quickly. The Reformed believed that their translation of the Hebrew and Greek to be inspired, literally God’s Word. The moderns believe that every translation is imperfect, and the Hebrew and Greek must be consulted at every turn. The Reformed believed that there were no places of corruption, meaning that every jot and every tittle had been preserved and made available. The moderns believe that some passages and words have indeed fallen away, yet what we have is “good enough” to ascertain all important doctrines. The pastor who believes the modern conclusions will preach in a manner which aligns with the modern textual scholars. The layman who believes the conclusions will read their Bible in a manner which aligns with the modern textual scholars.

While it may be true that this is an “open handed issue” as it were, it nevertheless has a significant impact on the Reformed and non-Reformed churches in the world. It changes preaching, reading, and practice. For that reason, this remains an important debate. Every pastor preaching through the Bible verse by verse will eventually have to side one way or another on this issue. Every Christian reading through the Bible will eventually have to do the same. Unfortunately, that debate is not easily settled, as it more or less comes down to answering the question, “Does modern evidence disprove the conclusions of the Reformed?” Those that answer, “Yes” will adopt the methods of Dan Wallace, Andrew Nasselli, Mark Ward, and Dirk Jongkind in Bible reading and preaching. Those that answer, “No” will retain the translation of the Reformed, say “Jehovah” instead of “Yahweh,” and so on. My point is this – whatever conclusion a believer lands on will impact the practice of their faith in a very noticeable way.

It is for this reason that this conversation continues to come up. You can be unaware of your brother’s doctrinal views on a number of issues, but not this one. It is inevitably going to come up, because it impacts the Bible translation you carry around, the words you use to pray, and the way a pastor preaches. It is unavoidable, and it is my hope that the modern day Reformed are at least more friendly to those of the Textus Receptus position than they have been in the past decade or so. It is my opinion that it is absolutely the position of the Reformed tradition, even if that tradition is considered wrong according to the modern perspective.

TurretinFan Attempts to Make an Argument

Introduction

On Monday, anonymous blogster TurretinFan published an article titled The “Stable Text” King James Version Argument which can be found here. The point of the article is to say that the “misleadingly labeled” Confessional Bibliology position is the same position that the KJV translators had to respond to, and is essentially the position of the Roman Catholics. I want my reader to notice that this article is about 10 sentences worth of assertions, and the rest citations which he claims prove his point. TurretinFan hasn’t actually made an argument in this article, nor has he addressed the material I have written. Let’s break down his assertions and see if we have anything to engage with from the article.

Confessional Bibliology is Misleadingly Labeled

In the first sentence of the article, TurretinFan asserts that the label “Confessional Bibliology” is a misleading label. He does not explain why or how it is misleading. He continues to say that Confessional Bibliology “seems to recognize the authority of the original languages” but “seems to conform to whatever Greek or Hebrew was followed by the translators/revisers of the King James Version.”

Apparently it is news to TurretinFan that the King James Version translators used the base text which we now call the Textus Receptus to create the King James Version translation. Again, no argument provided by our anonymous poster, simply an assertion as to what “seems” to be the case. In other words, there is nothing we can reasonably engage with here.

The Stable Text Argument is a Roman Catholic Argument

TurretinFan’s major blunder here is a common error that Critical Text advocates make, which is to confuse an argument for the text with an argument for the authority of a translation. He references Clement VIII who is making an argument for the stability of the Latin manuscripts, which are translations. This is a common argument made by James White, and not original to TurretinFan. This is not the argument I am making, nor have I ever made. Perhaps if TurretinFan understood the moniker “Confessional Bibliology,” he would know that the position I defend is simply the position of chapter 1 of the 17th century confessions, which state, “The Old Testament in Hebrew, (which was the native language of the people of God of old,) and the New Testament in Greek, (which, at the time of writing of it was most generally known to the nations,) being immediately inspired by God, and by his singular care and providence kept pure in all ages, are therefore authentical; so as in all controversies of religion, the Church is finally to appeal unto them. But, because these original tongues are not known to all the people of God, who have right unto an interest in the scriptures, and are commanded in the fear of God, to read and search them, therefore the are to be translated into the vulgar language of every nation unto which they come, that the word of God dwelling plentifully in all, they may worship him in an acceptable manner, and, through patience and comfort of the scriptures, may have hope” (Westminster Confession of Faith, 1.6).

TurretinFan ends by stating that my objective is to move from manuscripts to a single, stable text. This is arguably the strangest assertion he has made in the article, as this is simply what happened in time and space. Is he arguing that the church uses manuscripts today? Do Christians carry around various manuscripts to church? It is already the case that the form of the Bible has moved from manuscripts to printed texts and translations. This is true for both the Critical Text and TR position. The question is not whether we should migrate from manuscripts to printed texts, that has already occurred as a matter of history, not my opinion. The question is what those printed texts represent, and which of those printed texts and translations should be used by the people of God.

Interestingly enough, in his critique of me, TurretinFan has critiqued the nature of his preferred text. If my desire is a single, stable text, and this is the “wrong priority,” then he is really saying that the Critical Text is not a stable text, and “it is a mark of wisdom to revise when we discover errors.” TurretinFan has no basis for determining what an “error” is, however, because his preferred textual methodology cannot say what is an error. His methodology can only determine what is most likely the oldest reading, which is an evaluation that is agnostic to the idea of original or authentic. The concept of error requires the idea of verity or authenticity, which the Critical Text methodology does not claim to provide.

Turretin Vs TurretinFan

So if TurretinFan’s critique of my position is that I am wrong to desire a single, stable text, then it is clear that his position is the opposite. The only history repeating itself is for gainsayers such as TurretinFan to question the authenticity of Scripture, such as 1 John 5:7, Mark 16:9-20, and John 7:53-8:11. Ironically, Turretin himself agrees with me in my doctrine of Scripture.

“By the original texts, we do not mean the autographs written down by the hand of Moses, of the prophets and the apostles, which certainly do not now exist. We mean their apographs which are so called because they set forth to us the word of God in the very words of those who wrote under the immediate inspiration of the Holy Spirit” (Turretin, Francis. Institutes of Elenctic Theology. Volume 1. p. 106). In other words, Turretin rejects the idea that the Bible is inerrant in its autographs, but rather that it was kept pure in all ages via the copies or apographs.

Turretin continues by defining the difference between a textual variant and a corruption, “A corruption differs from a variant reading. We acknowledge that many variant readings occur both in the Old and New Testaments arising from a comparison of different manuscripts, but we deny corruption (at least corruption that is universal)” (Ibid., 111). In other words, Turretin does not agree with the modern critical text idea that there are variants which should be marked with a diamond, indicating corruption.

Turretin finishes by defending the passages that TurretinFan himself rejects, “There is no truth in the assertion that the Hebrew edition of the Old Testament and the Greek edition of the New Testament are said to be mutilated; nor can the arguments used by our opponents prove it. Not the history of the adulteress (Jn. 8:1-11), for although it is lacking in the Syriac version, it is found in all the Greek manuscripts. Not 1 Jn. 5:7, for although some formerly called it into question and heretics now do, yet all the Greek copies have it, as Sixtus Senesis acknowledges: “they have been the words of never-doubted truth, and contained in all the Greek copies from the very times of the apostles” (Bibliotheca sancta [1575], 2:298). Not Mk. 16 which may have been wanted in several copies in the time of Jerome (as he asserts); but now it occurs in all, even in the Syriac version, and is clearly necessary to complete the history of the resurrection of Christ” (Ibid., 115).

Here is the most entertaining conclusion that we can arrive at by actually reading Turretin:

  1. Turretin lots TurretinFan in with the heretics in his rejection of 1 John 5:7
  2. Turretin was aware of the manuscripts TurretinFan prefers (Jerome’s referenced manuscripts)
  3. Turretin details how doctrines are impacted by the removal of passages as found in the Critical Text (the history of the resurrection of Christ)
  4. Turretin rejects the notion of corruption (indeterminate readings, which are found in the ECM)
  5. Turretin rejects the reframing of the WCF by Warfield and the Chicago Statement

In other words, Turretin disagrees with TurretinFan on nearly every point as it pertains to Scripture, both textually and theologically. It is interesting that TurretinFan accuses me of making a Roman Catholic argument. This is a little known fact, but the Textus Receptus was actually the text of the Protestant Reformation. It makes no sense that defending such a text would be considered…Roman Catholic. I suspect the reason for such shoddy and inept argumentation is due to the fact that this is simply a regurgitation of assertions made by James White. Let me remind my reader that assertions are not arguments, and you actually have to explain why a quote supports your argument. This is persuasive writing 101. What we have to recognize is that a community college English 101 professor would send this article by TurretinFan back for revision because he hasn’t made an argument, he’s simply made assertions in between citations.

Conclusion

The greatest condemnation of this article is not it’s poor structure and lack of persuasion, it is the fact that Turretin himself disagrees with TurretinFan so fundamentally that he should really consider changing his moniker. Perhaps he should consider Dr. SpyOptics, the bargain bin version of Dr. Oakley.

Death of Logic

Welcome to the new year, reader. In 2024, I fully expect that we will find ourselves debating the same arguments and talking points afresh. The reality is that many of the theological controversies prevalent today are already resolved. Consider topics such as soteriology and eschatology. These have been discussed at great length by men much more capable than anybody alive today, yet modern men exhaust themselves, thinking that they have come up with some clever argument. Despite this, Christians will descend upon internet forums to cast their thoughts into the void, regardless of outcome. Such is the nature of debate disconnected from the real world.

Many, if not all, of the popular debates can be resolved by logical rules as simple as the law of noncontradiction or excluded middle. In other words, two things that cannot be true at the same time or one of two things must be true, given the claim requires it. The entire textual discussion can, at the very least, be reframed by evaluating foundational claims by these two laws of thought. For example, if our primary claim is “Scripture exists”, then we can exclude any claim which posits that “Scripture does not exist.” By “exist” we mean here that the Scriptures exist in such a way that we can access and use it. Many modern views of Scripture attempt to redefine the term “exist” to mean that the words are out in the world, somewhere, we just don’t have access to them. Yet, this is an illogical claim, for this view also allows for the possibility that the words are not out there in the world. This is why textual scholars hedge their claims with words such as “possible” and “likely.” In short, if it is “possible” that a text is original, it is also possible that a text is not original. Using this type of language violates the law of noncontradiction if we are trying to defend something which is said to “exist.” Something cannot exist and “possibly exist” at the same time.

Herein lies the nuance which muddles the debate. If you are a Christian, your premise is that Scripture exists, not that it “possibly exists.” This is a requirement for all Scriptural truth claims. We do not say in evangelism that, “It is possible that God so loved the world..” and so on. When we present the Gospel, we do so assuming the premise that the words exist, and that the words we have in front of us are correct. There is not a scenario where a Christian can confidently present the free offer of salvation given that the words on the page could be inauthentic. Yet, if you inspect the average claim of the modern bibliology camp, this is how they argue for Scripture. In one context, they read their Bible, listen to sermons, and debate theology with absolute confidence in the words on the page, and in the text-critical context, they argue with varying degrees of “possibility.” We know that this is not logically valid.

If we investigate what it means for the Bible to “exist,” we find that this concept is captured within the doctrine of preservation. To the traditional camp, the Bible exists because we have always had it, despite attempts at corruption. To the modern camp, the Bible exists insofar as we have it today. The first group has, at the very least, logical consistency because the definition of “Bible” has not changed over the years. The second group has a serious problem because in order to make the claim that the Bible exists, they must change the form of the Bible to mean something different in every age. The Bible, in other words, is more of a concept than an object. This is the primary difference between the two groups, as far as I can tell.

The traditional group sees the Bible as an object with a defined number of words, whereas the modern group sees the Bible as an object with a defined number of ideas. This is the explanatory mechanism the moderns use to describe how the words can be added, removed, or changed, and the thing (Bible) still be the same. According to this model, so as long as the core ideas are preserved, the Bible is preserved. This is a clever formulation, but unfortunately it has no explanatory mechanism. It cannot be demonstrated to be true, which means it is purely conjectural. This is due to the fact that literary ideas are derived from words. A text cannot preserve an idea without words. So if it were the case that a text can change while the ideas remain constant, one would have to actually demonstrate that additions, subtractions, and substitutions do not alter the substance of the text. This of course cannot be demonstrated to be true because any addition, subtraction, or substitution must, by definition, alter the substance of a text. This is how words work. Unless a modernist is willing to make the claim that all omitted portions of text are void of meaning, or all substitutions are merely synonymic, this must be the case. Ultimately, this claim is evaluated true on the simple merit that it has been said out loud.

We see that the textual debate is really a matter of definitions. In order for the modernists to be correct, they must employ a functional definition of “Bible” and “exists” and “preserve” in a different sense than the traditionalists. This is how they use historical writings such as the Westminster Confession to defend their claims. “If by Scripture they mean this, and if by preserved they mean this, then the Scriptures are indeed preserved!” Yet we can all agree that simply changing the definition of words does not win an argument, it simply means that the argument itself is different. This is my main point, the modern position of bibliology is comprised of claims which are substantially different than that of the traditionalists. That is to say, when these two groups collide in debate, they are using different functional definitions of key terms such as “Bible” and “preserved.”

This is important to recognize the next time you engage in a debate over a text such as 1 John 5:7. The textual modernist views his Bible as a collection of concepts and ideas, not a defined collection of texts. That is why a primary argument against the importance of the Comma Johanneum is that “the idea is contained elsewhere.” What I want my reader to understand is that this is actually an argument in support of a different definition of “Bible.” Take time to notice when somebody argues that, “1 John 5:7 isn’t even about the Trinity” or “It wasn’t even quoted at Nicaea in defense of the Trinity.” They are arguing this because in their mind the Bible isn’t defined by the texts contained within it, but the ideas. Yet we have already established that in literature, ideas are derived from words. Thus, adding, removing, or altering the words necessarily adds, removes, or alters the ideas.

Herein lies the primary logical problem with the modernist position on Scripture. In order for the text to mean something in itself, the ideas must be derived from the text. If it is the case that the ideas themselves are preserved apart from the text, then it is the case that the ideas have another preserving mechanism. Let us return to the argument related to the Trinity and 1 John 5:7. One of the chief arguments presented by textual modernists such as James White is that the doctrine of the Trinity can be derived elsewhere. Note the framing demands that the doctrine exists, and it can be found in the text. The doctrine is assumed first, and collected from Scripture as it exists to the modernist. In other words, the doctrine is preserved apart from the text. In this view, the reader brings a fully built out doctrine of the Trinity into Scripture. This is the implied assumption of this argument against the Comma Johanneum.

So we see the foundational logical problems that exist in the current iteration of the textual discussion. The modern side must redefine historical terms in order to fit their view into the Protestant frame. That is why some textual scholars avoid referring to Scripture as “The Bible” and instead opt for “bibles.” This is because at the core of the Critical Text position, the assumption is that “bibles” exist, but “the Bible” hasn’t existed since the first or second century. This is obviously problematic for Christian bibliology. The answer the modernists give is that, “we don’t need the Bible, as long as the ideas are preserved in the bibles.” One of the main problems in the textual discussion is that when a textual modernist employs the term “Bible” or “Scriptures,” they mean, “the bibles.” This demonstrates one of the foundational disconnects in the debate, both sides are using different functional definitions.

This is why the debate will continue into 2024 the same as it has since I have been involved. There are two theologies of the text, two logical foundations, two sets of definitions. My prediction is that 2024 will continue to highlight the differences in the positions with each discovery and evolution of the textual scholars. Happy New year and may the Lord bless you and keep you, reader.

Providential Exposure as it Relates to Preservation

The Theological Method and Preservation

The Theological Method for determining the text of Scripture heavily relies upon understanding the text that has been received by Christians, which is commonly called “exposure”. The text of Scripture is that which has been exposed to the people of God throughout the ages. John 10:27 says, “My sheep hear my voice, and I know them, and they follow me” (KJV). Michael Kruger, in his book Canon Revisited, says this,

“When people’s eyes are opened, they are struck by the divine qualities of Scripture – it’s beauty and efficacy – and recognize and embrace Scripture for what it is, the word of God. They realize that the voice of Scripture is the voice of the Shepherd” (101). 

This might seem like subjectivism, but this is the historic doctrine that has been recognized throughout the ages by the theologians of the faith, most notably John Calvin and Herman Bavinck. This doctrine is not to be confused with the Mormon doctrine of “burning in the bosom”, which has been done by men like James White. Many false doctrines are based on truth, and here the Christian must recognize that God’s Word is the means that He is speaking to His people in these last days, regardless of how that doctrine has been twisted by other systems. 

What distinguishes this doctrine from its Mormon counterpart is that this reception of the Scriptures by the people of God is not purely individualistic. The text of Scripture, as it has been handed down, exists ontologically, not just subjectively. There is a concrete shape of God’s Word that exists, and the people of God have had that Word in every generation. The text of Scripture must primarily be viewed as a covenantal document given to the covenant people of God for their use in all matters of faith and practice (LBCF 1.1). That does not mean that all those professing Christianity throughout the ages have agreed upon what belongs in Scripture, or that every Christian has had access to the whole of Scripture in every generation. In fact, there are a multitude of Christians that do not have access to God’s Word, either by circumstance, or by choice. 

It is important to take note of how the Apostolic church received the text of Holy Scripture to understand the doctrine of exposure as it relates to preservation. In the New Testament, there is never a case where Scripture is said to be a gift delivered to individual people. The Scriptures were always a corporate blessing to the covenant people of God (Acts 15:14; Titus 2:14; 1 Peter 2:9). In the testimony of Scripture itself, we can see that God delivers His Word to the people of God, not individual people of God. So the doctrine of exposure does not crumble due to certain individuals not having a copy of the Bible at all times. If this were the case, the fact that there are Christians who simply do not own a Bible would discredit this doctrine altogether. 

Despite the fact that the Canon is recognized in part by its corporate reception, this doctrine of providential exposure does not rest on ecclesiastical authority, as the papists might claim. There is no one single church which is responsible for giving authority to the text of Holy Scripture. In fact, no church could give authority to the Scriptures, they are authoritative in themselves (αυτοπιστος). Kruger explains this well, “The books received by the church inform our understanding of which books are canonical not because the church is infallible o because it created or constituted the canon, but because the church’s reception of these books is a natural and inevitable outworking of the self-authenticating nature of Scripture” (106).

It must be stated, that Kruger makes the distinction between the canon, and the text of the canon, which is the common thought amongst conservative scholarship. Upon examination of the theological method  however, there does not seem to be good reason to separate the two. If the doctrinal foundation of providential exposure demonstrates the “efficacy of the Shepherd’s voice to call”(106), then it follows that there must be a definitive voice that does the calling. The name of a canonical book is simply not efficacious to call sinners to repentance and faith. Simply listing off the canonical list is not the Gospel call. So the material that is providentially exposed to the people of God must also contain the substance which is effective unto life by the power of the Holy Spirit. God has not just preserved the book sleeves of the Bible, He has also preserved the words within those book sleeves. 

Since the Bible is self-authenticating, Christians cannot look to the totality or purity of its reception to determine which books or texts of the Bible should be received today. That is to say that because the majority of the Christian people do not accept one passage as authentic today, does not mean that it has not been properly received in the past, or that it is not ontologically a part of the canon. A passage of Scripture may not have been accepted as canonical by various groups throughout history, and this has indeed been the case, usually due to theological controversy. 

It is antithetical to the Theological method to say that the Scriptures are self-authenticating, but then also say that people must authenticate those Scriptures by a standard outside of the Scriptures themselves. Either the Scriptures are self-authenticating, or they are not. Which is why evidence are great tools to defend the Scriptures, but those evidences can never authenticate those Scriptures in themselves. It is problematic to say that God’s Word has been preserved, and kept pure in all ages, and then to immediately say that He has done so imperfectly, or has not fully exposed that Word to His people yet. 

The Theological method provides a framework that actually gives more weight to historical thought as opposed to modern thought. It disallows for a perspective that believes that the people of God had lost or added passages of Scripture, and that these texts need to be recovered or removed. It prevents certain theories that the text evolved, or that Christ’s divinity was developed over time. It especially rejects the idea that the original text of the New Testament was choppy, crude, and in places incoherent, and that scribes smoothed out the readings to make the text readable. In fact, it exposes those manuscripts that are choppy and missing parts to be of poor quality by assuming that the Holy Scriptures were inspired by the Holy Spirit rather than invented by ostensibly literate first century Jews. 

There may have been localities that corrupted the text (usually intentionally), but this does not represent the providential preservation that was taking place universally. The vast majority of textual variants are due to Scribal errors, but the significant variants were certainly an effort of revision. A Scribe simply wouldn’t have removed or added 12 verses by accident. A great example is the idiosyncratic Egyptian manuscripts uncovered in the 19th and 20th centuries, which tend to disagree with the general manuscript tradition in important variant units. These manuscripts have been given tremendous weight in the modern period due to shifting views of inspiration and preservation. 

If the Scriptures truly were inspired and preserved, then one should expect that the text did not evolve, or that the closest representative of those originals would be riddled with short, abrupt readings. One would expect that in every stage of copying, Scribal errors would be purged out, and that the true readings would persevere. In fact, this phenomenon can be observed in the vast majority of the extant New Testament manuscripts, which has unfortunately been described as Scribal interference, or smoothing out the text. When the transmission of the New Testament manuscripts are viewed Theologically, an entirely different story is told by the manuscripts which largely disagrees with the modern narrative which favors those choppy manuscripts which existed in one locality of the Christian world. 

The preservation of God’s Word can be demonstrated evidentially, but not without the proper Theological lens. Evidential arguments can be a powerful tool in all disciplines, but they often are not effective in themselves to change anybody’s mind. That is why the Theology of scholars will ultimately determine the manuscripts they deem to be earliest and best. Simply counting manuscripts, or weighing manuscripts, is simply not consistent with the conservative doctrine of preservation. In both cases, these methods attempt to take an external authority, such as manuscript count, or the age of the manuscript, and use that to authenticate the Word of God. Yet both of these are at odds with the doctrinal standard that is laid forth in Scriptures themselves, that the Word of God is self-authenticating. That is why the language of “the text that has been received” is warranted in this conversation, because it recognizes God’s providential preservation and exposure of the ontological canon to the people of God in every age. 

Conclusion 

The doctrine of exposure is often misunderstood as being too similar to the Mormon doctrine of “burning of the bosom” or the papal doctrine which states that Rome has the authority to authenticate the Bible. Despite these abuses of Scripture, the fact remains that the Scriptures are self-authenticating. It is easy to fall back onto empirical approaches, because they seem to be the most logical. Yet these empirical approaches do not do what they claim they can do, and this is becoming increasingly evident with each passing year. The number of Bibles has only increased, and exponentially at that. Modern methodology has not narrowed the text of the New Testament to fewer legitimate readings, but has expanded greatly the number of readings that “could be” original or early. 

The efforts of modern textual scholarship has only increased the uncertainty of the text of the New Testament. This has culminated in the abandonment of the search for the original text of the Bible for the Ausgangstext, or the earliest text scholars can get back to (which is 3rd or 4th century). Practically speaking, this pursuit will simply result in arriving at some hypothetical form of the text that may have existed in Egypt in the third century. Since this seems to be the direction of most current New Testament text-critical scholarship, it seems that it is time to return to the old paths. The Theological method has been expressed by countless Theologians of the Christian faith, and it should not be abandoned for the sake of adopting the modern critical scientific method. The Scriptures should always be handled as self-authenticating, and a shift to this way of thinking would result in a massive change in the direction of modern New Testament scholarship.